Introduction

The integration of Artificial Intelligence (AI) into various sectors of society has become a prominent focal point of academic inquiry and application, with education standing out as a particularly noteworthy arena (Nazaretsky et al., 2022a). Historical traces of AI's foray into education can be linked to the development of early chatbots in the 1960s (Weizenbaum, 1966). However, a marked improvement in AI's capabilities, particularly in the realm of generative AI, became apparent in last decade (Adiwardana et al., 2020; Vaswani et al., 2017). Among the torchbearers of this evolution is ChatGPT, an application of Generative AI with transformative potential for educational landscapes (OpenAI, 2023).

Generative AI, such as ChatGPT, operates distinctively from traditional AI models. Instead of merely recognizing patterns and making predictions, it innovatively produces novel outputs, encompassing text, audio, and video. With its roots in deep neural network architectures, and further powered by transformer designs, Large Language Models (LLMs) like ChatGPT or LLaMA harness billions of parameters, making strides in various natural language processing tasks (OpenAI, 2023; Touvron et al., 2023). Notably, educators have leveraged the capabilities of such LLMs for diverse purposes, ranging from augmenting learning experiences through dynamic feedback to the facilitation of language acquisition (Cukurova et al., 2023).

However, the infusion of such AI technologies in post-secondary education is not without challenges. One can point to the surge of interest in AI's educational potential over the last decade, juxtaposed with a landscape marked by exaggerated claims and often inconclusive findings (Bates et al., 2020). Even as a considerable volume of studies on AI in education emerged in the past years, signifying a booming interest, the actual effects of applications like ChatGPT on higher education institutions and stakeholders remain enigmatic (Zawacki-Richter et al., 2019).

Moreover, accompanying this technological fervor are grave concerns about intellectual property, data privacy, potential biases, and equitable access to resources (Li et al., 2022). The deployment of generative AI in pedagogical contexts begs pertinent questions about pedagogy, learning paradigms, authorship, and more (Williamson et al., 2023; Zembylas, 2023). For instance, as students turn to AI for academic aid, intricate issues of plagiarism, erasure of original thought, and challenges to established pedagogies arise (Rudolph et al., 2023). Such developments resonate with broader ethical considerations surrounding factuality, fairness, transparency, and more, emphasizing the need for sustainable AI systems in education (Kasneci et al., 2023). Consequently, there is an urgent and pressing need to explore not just the promises but also the potential impacts, limitations, and ethical ramifications of LLM-enabled educational applications (Strzelecki, 2023). Research must tackle the intricacies of design and implementation, ascertain the appropriateness of such AI deployments in educational settings, and articulate the overarching principles and guidelines for ethical utilization (Hahn et al., 2021).

This investigation, therefore, situates itself within this rich tapestry of debates, possibilities, and challenges. By delving deeper into the matrix of technology adoption, artificial intelligence, and the nuances presented by applications like ChatGPT, we aim to contribute to a more informed, equitable, and constructive discourse that respects both the transformative potential of AI and the foundational tenets of education in the twenty-first century.

ChatGPT, a chatbot tool, uses artificial intelligence to facilitate human-like conversations (Vaswani et al., 2017). This innovative natural language processing tool is capable of providing assistance with various tasks, including email composition, essay writing, and coding. OpenAI developed ChatGPT, which was officially launched on November 30, 2022, and has since attracted a great deal of attention (OpenAI, 2023).

Despite its impressive capabilities, ChatGPT is currently in research and feedback- gathering phase, and it is freely accessible to the public for usage. However, a paid subscription version called ChatGPT Plus was launched on February 1st, 2023, offering improved response times, priority access to new improvements and features. Milmo (2023) reports that ChatGPT has become the fastest-growing application in history, with approximately 100 million users actively engaged within two months of its release, as per Forbes. Along with the ability to answer basic queries, the model offers a variety of functions, including generating AI art prompts, coding, and crafting essays. Both supervised and reinforcement learning techniques are employed to fine-tune the language model, with human AI trainers participating in conversations where they perform the roles of both user and AI assistant. Despite its many strengths, ChatGPT has limitations such as difficulty answering questions that have a specific wording and a lack of quality control, which can lead to erroneous responses (Altman, 2023).

ChatGPT in university education presents challenges to the academic community wherever ChatGPT is available (Teubner et al., 2023). While some academics embrace the development, others oppose its use for fear that it may facilitate cheating or academic misconduct. Proponents argue that adapting to the technology and designing policies for academic integrity will equip students with the necessary skills and ethical mindset to operate in an AI environment. The chatbot generates coherent and informative text, and can be accessed in many countries (Liebrenz et al., 2023). Academics are at loggerheads over the use of AI-generated bots in educational settings. Some academics promote a framework for its appropriate use, whereas others, like Hsu (2023), advocate an outright ban. The escalating debate underscores a need for reshaping curricula in higher education to embrace more critical thinking, ethical perspectives, creativity, problem-solving, and future-proof solutions, preparing students for an AI-influenced professional landscape.

Though ChatGPT has swiftly gained popularity as a cutting-edge tool in natural language processing, research into how it's perceived and utilized by university students remains sparse. This shortfall is significant because discerning how students engage with AI aids is crucial for their efficacious deployment in educational contexts. Consequently, a deeper understanding is needed regarding factors that shape students' attitudes towards ChatGPT, their perceptions of its usefulness, and the difficulties they encounter while using it. By closing this research gap, we can gain a clearer perspective on how best to weave ChatGPT into the fabric of our education system, facilitating student learning and academic triumph.

This paper seeks to fill that research void by examining the acceptance and application of ChatGPT among university students. We delve into the determinants that foster or inhibit its acceptance and utilization, utilizing the "Unified Theory of Acceptance and Use of Technology 2" (UTAUT2) model by Venkatesh et al. (2012) as our theoretical framework. We've expanded this model by integrating "Personal Innovativeness"—a construct introduced by Agarwal and Prasad (1998) in the IT domain, viewing it as a potential precursor within the model.

We begin by outlining the origins of ChatGPT and summarizing the controversy surrounding its use. Then we proceed to explain how we've extended the UTAUT2 model to evaluate students' acceptance and use of ChatGPT. A specifically modified measurement scale tailored for ChatGPT in university settings is also presented. Following this, we share the results of our analysis, deploying a structural equation modeling using the partial least squares method and the model estimation. In this section, we also engage in a rich discussion of our findings, highlighting the uniqueness and contributions of our study. We wrap up by addressing the limitations of our study and recommending avenues for future exploration.

Literature Review

The literature indicates a burgeoning interest in the application of ChatGPT within higher education, spanning general, and scientific educational domains. Studies highlight its potential in personalizing learning and enhancing engagement but caution against uncritical adoption due to biases and academic integrity concerns (Kasneci et al., 2023; Rudolph et al., 2023). Cotton et al. (2023) and Tlili et al. (2023) discuss maintaining academic integrity, while Williamson et al. (2023) encourage more research into AI's educational impact. Naumowa (2023) stresses critical thinking in using AI, whereas Hsu (2023), Ivanov and Soliman (2023), Lim et al. (2023), Crawford et al. (2023), Sullivan et al. (2023), and Cooper (2023), explore the ethical and practical implications, calling for ethical guidelines and theoretical frameworks.

The role of ChatGPT in academic authorship raises concerns about integrity and recognition of contribution. Van Dis et al. (2023) recognize its benefits in democratizing knowledge but call for debates on authorship and accountability. The potential impact of AI on academia and the need for ethical use policies is discussed by Lund et al. (2023), while Thorp (2023) cautions against naming AI as a co-author. Perkins (2023) advises HEIs to update academic integrity policies to consider AI tools. Conversations with ChatGPT reveal its capabilities and limitations, particularly in areas like plagiarism and misuse in education (Eysenbach, 2023). Interviews explore its potential in improving content creation and academic research, with a focus on ethical implications (Lund & Wang, 2023; Pavlik, 2023). These dialogues with ChatGPT underscore the importance of evaluating its impact on academia and the media critically.

Methodology

The acceptance and use of technology has become a significant topic of interest in recent years, as individuals increasingly rely on technology for daily tasks. One of the most prominent theories used to explain and predict technology acceptance is the “Unified Theory of Acceptance and Use of Technology” (UTAUT) (Venkatesh et al., 2003). UTAUT “synthesizes the concepts and user experiences that provide the foundation for theories on the user acceptance process of an information system” (Yu et al., 2021). UTAUT was constructed by integrating and synthesizing eight pre-existing models of information technology acceptance. The model includes constructs such as "Performance expectancy”, “Social influence”, "Effort expectancy”, and Facilitating conditions”, which have been found to significantly influence "Behavioral intention” to use a technology. Additionally, individual differences such as age, gender, voluntariness of use and experience are considered moderators of the four constructs in the UTAUT model.

Recently, UTAUT was modified to create UTAUT2, which includes three new constructs of “Hedonic motivation”, “Price value”, and “Habit” (Venkatesh et al., 2012). UTAUT2 is a prominent theoretical model that aims to understand the factors that influence individuals' adoption and use of new technologies in organizational and personal contexts (Tamilmani et al., 2021). Developed through extensive empirical research, UTAUT2 provides a comprehensive framework for researchers and practitioners to identify key factors that influence the acceptance and use of technology. In the realm of higher education the UTAUT2 model is used to identify factors affecting students' or teachers' intentions to use different technology tools such as e-learning systems (Raza et al., 2022), mobile applications (Ameri et al., 2020; Kang et al., 2015), immersive virtual and augmented reality (Bower et al., 2020; Faqih & Jaradat, 2021), and learning management software (Alotumi, 2022; Jakkaew & Hemrungrote, 2017; Kumar & Bervell, 2019). Recent studies also examine how these factors can be affected by different contextual factors, such as the COVID-19 pandemic (Osei et al., 2022; Zacharis & Nikolopoulou, 2022), and how they can be utilized to improve the design and implementation of educational technology tools.

Our study argues that “Performance expectancy”, “Effort expectancy”, "Social influence”, "Facilitating conditions”, “Hedonic motivation”, Price value” and “Habit” can significantly influence students’ “Behavioral intention” to use a ChatGPT technology in higher education. We proposed to extend a well-established UTAUT2 theory, by adding also “Personal innovativeness” as a factor influencing “Behavioral intention” to use ChatGPT technology. Personal innovativeness (PI) has been acknowledged as important factor that impact technology adoption and usage (Agarwal & Prasad, 1998). A number of scholars “have suggested that personality traits, such as personal innovativeness, play a significant role in technology adoption in the domain of Information Technology (IT)” (Farooq et al., 2017; Sitar-Taut & Mican, 2021). Personal innovativeness reflects a person’s tendency to independently experiment with and implement new IT developments, and is seen as a stable and context-specific characteristic that strongly influences acceptance and adoption of IT (Dajani & Abu Hegleh, 2019; Twum et al., 2022).

The goal of our research is to examine the influence of seven UTAUT2 factors and PI on students' behavioral intention to use ChatGPT and its effectiveness in assisting them with their studies. The study will define each factor and explore how students' perceptions of ChatGPT's use influence their continued usage.

Performance expectancy (PE) “influences individuals' behavioral intention to adopt new technology. It refers to the extent to which individuals believe that using a system will help them attain gains in job performance or enhance their performance in learning processes” (Venkatesh et al., 2003). In this study PE pertains to how much these students think that utilizing ChatGPT can boost their academic results or efficiency. We hypothesize that:

  • H1: “Performance expectancy will have a direct positive influence on the Behavioral intention to use ChatGPT in higher education by students.”

Effort expectancy (EE) is defined “as the degree of ease or effort associated with the use of technology” (Venkatesh et al., 2003). EE consists of constructs such as perceived ease of use, complexity, and ease of use. Research has shown that EE is a critical predictor of technology acceptance, and it has a direct impact on individuals' “Behavioral intention” to use technology. In this study, EE would denote how much students feel that interacting with ChatGPT is easy and demands minimal effort. Hence, it is proposed that:

  • H2: “Effort Expectancy will have a direct positive influence on the Behavioral intention to use ChatGPT in higher education by students.”

Social influence (SI) is defined “as the extent to which important others, such as family and friends, believe that an individual should use a particular technology” (Venkatesh et al., 2003). The influence of social circles, including family members, teachers, co-workers, elders, friends, and peers, has been shown to positively impact users' intention to use a technology. In this research, SI indicates the extent to which students perceive that peers, instructors, or other key figures in their social circle endorse or motivate them to engage with ChatGPT. We hypothesize that:

  • H3: “Social influence will have a direct positive influence on the Behavioral intention to use ChatGPT in higher education by students.”

Facilitating conditions (FC) refer to “the level of accessibility to resources and support needed to accomplish a task” (Venkatesh et al., 2003). In university education settings, FC emphasize the importance of having access to technical and reliable infrastructure, knowledge, training, and support, which can impact students' willingness to use educational systems. FC in this context means the extent to which students feel they can access the AI tool even with its popularity, coupled with their access to technical assistance and training for ChatGPT. The subsequent hypotheses are stated:

  • H4: “Facilitating conditions will have a direct positive influence on the Behavioral intention to use ChatGPT in higher education by students.”

  • H5: “Facilitating conditions will have a direct positive influence on the ChatGPT Use behavior in higher education by students.”

Hedonic motivation (HM) refers “to the pleasure or enjoyment derived from using a technology, and its effect on the intention of users has been found to be statistically significant” (Venkatesh et al., 2012). The findings of various studies suggest that users are more likely to continue using a technology if they experience pleasure or enjoyment while using it. In this context, HM denotes how much students perceive ChatGPT as fun or pleasurable to engage with, and their enjoyment in exploring new AI technological tools. Hence, it is proposed that:

  • H6: “Hedonic motivation will have a direct positive influence on the Behavioral intention to use ChatGPT in higher education by students.”

Price value (PV) is “an individual's trade-off between the perceived benefits of using the system and its monetary cost” (Venkatesh et al., 2012). PV refers to the cost students bear when purchasing access to the online services used for learning. Hence, it is proposed that:

  • H7: “Price value will have a direct positive influence on the Behavioral intention to use ChatGPT in higher education by students.”

Habit (HT) is defined “as the extent to which an individual tends to perform behaviors automatically because of prior learning and experiences with the technology” (Venkatesh et al., 2012). Venkatesh et al. (2012) and Limayem et al. (2007) have explained HT as a perceptual construct, and it has been found to be a significant predictor of “Behavioral intention” and technology use (Tamilmani et al., 2019). In the context of the study, HT pertains to how deeply students have ingrained the use of ChatGPT into their academic habits. This encompasses aspects like how often they use it, how long they engage with it, and how seamlessly ChatGPT fits into their academic processes. The subsequent hypotheses are stated:

  • H8: “Habit will have a direct positive influence on the Behavioral intention to use ChatGPT in higher education by students.”

  • H9: “Habit will have a direct positive influence on the ChatGPT Use behavior in higher education by students.”

The connection between Personal Innovativeness (PI) and acceptance and usage of technology has been thoroughly investigated in the literature (Slade et al., 2015). This characteristic can be interpreted “as a willingness to adopt the latest technological gadgets or a propensity for risk-taking, which may be associated with trying out new features and advancements in the IT domain” (Farooq et al., 2017). Within the scope of the study, personal innovativeness relates to the degree of eagerness students show in adopting cutting-edge tech tools such as ChatGPT, as well as their belief in their capacity to learn and excel in new technological skills.

  • H10: “Personal innovativeness will have a direct positive influence on the Behavioral intention to use ChatGPT in higher education by students.”

Behavioral intention (BI) is a fundamental concept in the study of technology adoption and use behavior (Ajzen, 1991; Davis, 1986; Fishbein & Ajzen, 1975). It refers to “individuals' willingness and intention to use a particular technology for a specific task or purpose” (Venkatesh et al., 2003, 2012). In the context of this study, BI is investigated to understand students' willingness to use ChatGPT in their studies.

  • H11: “Behavioral intention to use will have a direct positive influence on the ChatGPT Use behavior in higher education by students.”

Use behavior (UB) aims to explain user technology acceptance and usage behavior. According to the model, actual use refers to the extent to which users actually use a technology for a specific task or purpose. The original study by Venkatesh et al. (2012) did not provided how the actual use was measured. In this study use of ChatGPT was measured on seven items scale from “never” to “several times a day”.

Venkatesh et al. (2016) recommended using UTAUT2 as a fundamental model to establish hypotheses regarding the connections between proposed variables and user adoption of technology. Additionally, Dwivedi et al. (2019) noted that previous research in this field typically utilized only a portion of the UTAUT2 model and frequently neglected moderators. To address this issue, the current study employed a modified version of UTAUT2 and did not incorporate age, gender, and experience as moderators.

Our study utilized a research model (Fig. 1) which included seven constructs from the UTAUT2 scale (Venkatesh et al., 2012). This scale is commonly employed to evaluate the acceptance of technology. To further enhance our model, we integrated the PI variable from Agarwal and Prasad's (1998) research. The data was collected using a seven-point Likert scale, where respondents could choose from "strongly disagree" to "strongly agree". The scale for measuring use behavior was 7-options from “never” to “several times a day”. and Descriptive statistics and measurement scale are in Appendix Table 3.

Fig. 1
figure 1

An extended UTAUT2 model for ChatGPT – proposition

Sample Characteristics

First, a pilot study was carried out using a representative sample of 36 students (18 female and 18 male) from the first cycle study level to test the scales developed for this study before administering them to the targeted audience. The reliability and validity criteria were met for each construct, and discriminant validity was confirmed (Hair et al., 2013). In terms of determining the final sample size, according to Hair et al. (2022), studies conducted with the PLS-SEM method require a minimum sample size of 189 to detect R2 values of at least 0.1 with a significance level of 5%. Moreover, Arnold (1990) suggested that a statistical power of at least 95% is commonly sought after in social science research.

As of December 31, 2021, there were 1,218,200 students in universities in Poland. To calculate the sample size of population size 1,218,200, with confidence level at 95% and margin of error 5%, we use the following formula provided by Yamane (1967) “n = (z^2 * p * (1-p)) / e^2, where n is the sample size, z is the z-score associated with the confidence level, which is 1.96 for a 95% confidence level, p is the estimated proportion of the population with the desired characteristic”. Since we don't have an estimate, we use 0.5 to get the maximum sample size and e is the margin of error, which is 0.05. Plugging in the values we get a minimum sample size of 385.

The questionnaire was distributed through a web survey created on Google Forms. A total of 13,388 students from the University of Economics in Katowice, Poland, were invited to participate via email in the period of 13th to 21st March 2023. To ensure unbiased data, the students were assured of the confidentiality, voluntary participation, and anonymity of their responses. After removing 3 responses with a variance of zero, 503 valid responses were collected, which is enough, according to a minimal sample size calculation.

The sample size consists of 268 males (53.3%), 212 females (42.1%), and 23 (4.6%) who were unwilling to disclose their gender. With regard to the diversification of study experience, 25 students (5%) are in their first year, 162 (32.2%) are in their second year, and 192 (38.2%) are in their third year of the bachelor's degree program. Additionally, 52 students (10.3%) are in their first year and 64 (12.7%) are in their second year of the master's degree program, while 8 (1,6%) are PhD students.

Results

We implemented the PLS-SEM technique using the path weighting approach via SmartPLS 4 software (Version 4.0.9.2), adopting the standard initial weights with a limit of 3000 iterations to estimate our model (Ringle et al., 2022). Additionally, to determine the statistical significance of the PLS-SEM outcomes, we applied bootstrapping with 5000 samples, which is a nonparametric statistical method. In the assessment of constructs specified reflectively, we scrutinized the indicator loadings, noting that a loading greater than 0.7 indicates the construct explains over half of the indicator’s variance, signifying an acceptable degree of item reliability. Loadings are presented in Appendix Table 3 and exceed the lower bond except the FC4, which was removed from additional processing within the model and is not taken into account.

Composite reliability is a criterion for evaluating reliability, where results from 0.70 to 0.95 demonstrate reliability levels that are acceptable to good (Hair et al., 2022). Cronbach’s alpha measures internal consistency reliability and uses similar thresholds as composite reliability (ρc). An additional reliability coefficient ρA is built on Dijkstra and Henseler (2015), with ρA providing an exact and consistent alternative. The assessment of the measurement models' convergent validity is achieved by calculating the average variance extracted (AVE) from all the items linked to a specific reflective variable. An AVE threshold of 0.50 or higher is considered acceptable (Sarstedt et al., 2022). All measurement instruments have met the quality criteria (Table 1).

Table 1 Model measurement instruments

The Henseler et al. (2015) heterotrait-monotrait ratio of correlations (HTMT) is the preferred method for PLS-SEM discriminant validity analysis. A threshold of 0.90 is recommended since a high HTMT result suggests a difficulty with discriminant validity, when the constructs are conceptually similar. If they are more distinct, a limit of 0.85 should be used (Henseler et al., 2015). All the values in are below a 0.85 threshold (Appendix Table 4).

The subsequent stage involves an assessment of the coefficient of determination (R2), which gauges the amount of variance accounted for in each construct and hence the model's explanatory capability. The coefficient of determination, denoted as R2, varies from 0 to 1, with higher figures denoting a stronger degree of explained variance. According to Hair et al. (2011), R2 values can generally be interpreted as weak (0.25), moderate (0.50), or substantial (0.75). When evaluating the magnitude of an independent variable's impact, f2 values are categorized as large (0.35), medium (0.15), or small (0.02), and are used as standard reference points. Should the effect size fall below 0.02, it is indicative of a negligible effect (Sarstedt et al., 2022).

PLS-SEM findings are illustrated in Fig. 2, with standardized regression coefficients shown on the path relationships and R2 values presented in the variables' squares. The primary observation reveals that HT has the most prominent impact (0.339) on BI, followed by PE (0.260), and HM (0.187), explaining 72.8% of the "Behavioral intention" variance (as indicated by the R2 value). SI (0.093), PI (0.086), PV (0.083), and Effort expectancy (0.079) have positive effect on Behavioral intention, however there is no f2 effect size for these relationships. Conversely, BI has the most significant effect (0.424) on UB, with HT (0.255) and FC (0.188) following suit. These three variables account for 54.7% of the "Use behavior" variance. Table 2 presents results of the significance testing for the path coefficients within the structural model and the verification of the hypotheses.

Fig. 2
figure 2

An extended UTAUT2 model for ChatGPT – results

Table 2 Path coefficients in the structural model and the results of the significance tests

Discussion

Our study enhances the knowledge around student perceptions of ChatGPT's application in tertiary education. Although existing literature on this subject is scarce, particularly regarding higher education, our results hold considerable importance for broadening discussions about the deployment of AI-driven chat tools in academic settings. We utilized the UTAUT2 scale to evaluate the acceptance and utilization of ChatGPT, with all eight external variables satisfying the established standards for reliability and validity. Our study confirms that three UTAUT2 constructs: “Performance expectancy”, “Habit”, and “Hedonic motivation” are positively connected to “Behavioral intention”, consistent with Zwain (2019) study on users' acceptance of learning management systems.

Our research affirms the strong association between “Performance expectancy” and “Habit”, which is consistent with previous studies by Venkatesh et al. (2012) and Lewis et al. (2013) on how emerging technologies are accepted in higher education. Within the framework of ChatGPT, an AI-powered conversation platform, students demonstrate a high level of comfort in adopting new technology, and the frequency of their usage contributes to the development of routine actions. “Habit” was found to have a significant positive impact on “Behavioral intention” in a majority of UTAUT2-based studies in higher education. For instance, it has been shown to positively influence acceptance and use of lecture caption systems (Farooq et al., 2017), mobile learning during social distancing (Sitar-Taut & Mican, 2021), and Google Classroom platform for mobile learning (Alotumi, 2022; Jakkaew & Hemrungrote, 2017; Kumar & Bervell, 2019). Nevertheless, our results are not the same as those of Twum et al. (2022), who found no direct effect of “Habit” on “Behavioral intention” to use e-learning, and Raman & Don (2013) and Ain et al. (2016), who found no direct impact of “Habit” on “Behavioral intention” to use learning management systems. Consistent with the original UTAUT2 model, our study confirms the significant relationship between Habit and Use behavior.

In our study, “Performance expectancy” emerged as the second-best indicator of “Behavioral intention”. These findings are consistent with other studies that have demonstrated a notable and affirmative correlation between “Performance expectancy” and “Behavioral intention” in areas such as mobile learning (Arain et al., 2019), e-learning platforms (Azizi et al., 2020; Nikolopoulou et al., 2020), and learning management software (Raman & Don, 2013). In addition, “Performance expectancy”, as an original variable in the UTAUT theory, has been shown to be a significant predictor of “Behavioral intention” in studies such as Raffaghelli et al.'s (2022) exploration of students' acceptance of an early warning system in higher education and Mehta et al.'s (2019) investigation of the influence of values on e-learning adoption. According to our findings, students with high "Performance expectancy" are more likely to adopt useful technology like ChatGPT.

Our study found that “Hedonic motivation” has a positive impact on “Behavioral intention” to use ChatGPT. This finding is in line with earlier studies on the use of emerging technologies in higher education, including Google Classroom (Kumar & Bervell, 2019), animation (Dajani & Abu Hegleh, 2019), mobile technologies (Hu et al., 2020), blended learning (Azizi et al., 2020), and augmented reality technology (Faqih & Jaradat, 2021). However, our findings contradict those of studies that examined the use of learning management systems (Ain et al., 2016) and blackboard learning systems (Raza et al., 2022).

“Effort expectancy”, “Social influence” and “Price value” have been found to have a positive and significant effect on “Behavioral intention”, despite having f2 values that are lower than 0.02. With respect to “Effort expectancy”, students' responses had the highest mean values among all the variables, indicating that the AI-powered chat technology is widely used and there are no problems when using ChatGPT. This demonstrates that using this technology in higher education requires very little effort and does not affect “Behavioral intention”. Research has shown comparable outcomes on the application of mobile learning while social distancing (Sitar-Taut & Mican, 2021), using Google Classroom (Alotumi, 2022; Kumar & Bervell, 2019), mobile learning (Ameri et al., 2020; Arain et al., 2019; Kang et al., 2015; Nikolopoulou et al., 2020) e-learning platform (El-Masri & Tarhini, 2017; Twum et al., 2022) and learning management system (Ain et al., 2016; Zwain, 2019).

“Social influence” has a very low impact on the “Behavioral intention” to use ChatGPT. The likelihood of adopting the AI-enabled chat is higher among early adopters who have a higher level of education and are less swayed by external factors. Our study found no social pressure on the adoption of ChatGPT, possibly due to it being a new technology not yet widely adopted. Nonetheless, as academic institutions develop regulations regarding AI tools such as ChatGPT, the factor of "Social Influence" might gain significance. Previous studies have found “Social influence” to have a significant effect on “Behavioral intention” in areas such as mobile learning (Kang et al., 2015) and e-learning systems (El-Masri & Tarhini, 2017), while others did not find such significance, such as in the study of Alotumi (2022) or Kumar & Bervell (2019) on Google Classroom acceptance.

Likewise, “Price value” has a minimal effect on the “Behavioral intention” to use ChatGPT. This may be because the use of ChatGPT is currently free of charge during the tool's research preview. Although the ChatGPT plus version with the latest GPT-4 language model is available for $20 per month, the difference in the ChatGPT's use, powered by the GPT-3.5 version, for everyday student tasks is unlikely to be noticeable. However, students may perceive “Price value” as learning value since they need to invest time and effort into learning new technology. In the future, ChatGPT may no longer be offered as a free preview, and “Price value” could have a more significant impact. Previous studies have found “Price value” to have a significant effect on “Behavioral intention” in areas such as lecture caption systems (Farooq et al., 2017) and blended learning (Azizi et al., 2020), while others did not find this significance in areas like mobile learning (Nikolopoulou et al., 2020) and e-learning (Osei et al., 2022).

From the UTAUT2 model, the last construct, “Facilitating conditions”, has been found to have no significant effect on “Behavioral intention, but it did have a significant effect on “Use behavior”, as the original UTAUT2 model contains this relationship. Previous studies on the adoption of animation (Dajani & Abu Hegleh, 2019) and Google Classroom in higher education (Alotumi, 2022; Kumar & Bervell, 2019) have usually found that “Facilitating conditions” have no effect. On the contrary, it was found that “Facilitating conditions” have a significant impact on “Behavioral intention” in areas such as mobile learning (Yu et al., 2021) and augmented reality technology in higher education (Faqih & Jaradat, 2021).

“Personal innovativeness” is the fourth variable that has been found to have a minimal positive effect on “Behavioral intention”, due to f2 lower than 0.02. This finding could be since students have not fully experimented with ChatGPT and may have insufficient knowledge on how to use it. In previous studies, personal innovativeness was found to have a significant effect on “Behavioral intention” in areas such as animation (Dajani & Abu Hegleh, 2019), lecture capture systems (Farooq et al., 2017), mobile learning during social distancing in higher education (Sitar-Taut & Mican, 2021), and e-learning (Twum et al., 2022).

In this model, “Behavioral intention” is explained by 72.8%, which is considered a substantial explanatory power. Moreover, “Behavioral intention” has a direct and significant impact on “Use behavior” and model explains “Use behavior” by 54.7%, which is considered a moderate explanatory power. The explained variances were higher than in a study conducted by Maican et al. (2019) on communication and collaboration applications in academic environments, as well as in a study conducted by Hoi (2020) on how the mobile devices are accepted and used for language learning by students.

Recent research delves into the increasing discourse around the integration and application of AI-driven tools in education. There are suggested initiatives to bolster educators' proficiency in navigating multiple AI-enhanced educational platforms. For example, educators have opportunities to undergo specialized training modules to familiarize themselves with platforms that assess student essays (Nazaretsky et al., 2022a). Similarly, there are specific platforms tailored for educators that can be introduced in educational settings to elevate the learning experience (Cukurova et al., 2023). There are also tools designed to gauge the degree to which educators embrace AI-driven utilities and address potential issues like trustworthiness, the lack of human-like attributes, and clarity in AI's decision-making patterns (Nazaretsky et al., 2022b).

Our investigation pivots to a different facet of the educational journey, centering on students' acceptance and utilization of generative AI utilities. This research's value lies in shedding light on determinants shaping the embrace of such AI-driven tools. It's pertinent to mention that ChatGPT doesn't have adaptive capabilities and isn't tailored for academic use, indicating a need for more research on its potential role in learning environments.

Practical Implications

ChatGPT might assist with personalized learning and the adaptation of AI in education. For students, the role of ChatGPT can be highlighted as an efficient and tailored tool for information acquisition, conceptual understanding, language learning, and task facilitation. It can be perceived as a digital assistant capable of providing immediate, personalized feedback and guidance. It can serve as a grammar checker, content recommender, and query solver across a spectrum of disciplines and languages. ChatGPT's ability to present complex concepts in simpler terms for better comprehension can also be highlighted, along with its potential to make abstract educational materials more accessible. Additionally, it can has a role as a research facilitator, providing swift access to relevant information and a strategic roadmap for research studies.

The transformative potential of ChatGPT for learning, indicating its capability to create tailored content, interactive learning experiences, and customized assessments can be stressed, thereby offering a more individualized learning pathway. The use of AI technologies such as ChatGPT to supplement traditional educational strategies could result in a dynamic learning ecosystem that caters to each learner's unique needs and learning styles. It can provide individualized attention that human instructors, due to time and resource constraints, may not be able to. There can be underscored the potential of ChatGPT as a creative tool to stimulate critical thinking, enhance metacognition, and foster innovative problem-solving skills among students. By aiding in tasks such as organizing and summarizing information, developing essays, and creating tests based on study material, ChatGPT could be instrumental in promoting deeper learning.

However, there should be issued a note of caution about the ethical implications, data security, and potential overreliance on AI technologies. This emphasizes the importance of using AI as a tool, not as a replacement for human teachers, indicating that the criticality of human judgment and contextual understanding must remain integral to the educational process. The integration of AI technologies like ChatGPT in education, when done responsibly and ethically, can create enriched, personalized, and adaptive learning experiences. However, it requires concerted effort to ensure its appropriate use, involving iterative feedback loops between ChatGPT, human instructors, and students, along with careful attention to the potential challenges and pitfalls.

Limitations

The questionnaire was circulated during a turbulent phase related to AI, with each month profoundly influencing the tool's evolution. Those who began using ChatGPT early in its rollout are viewed as pioneers. Additionally, factors like specific times in the academic year, like assignment submissions or the specific subject of study, might affect how the tool is used. The study is also limited by the sample selection, which consisted only of students from one state university in Poland. However, the sample was diverse in terms of study experience. We did not consider different study programs as a moderating variable. It is possible that different study programs, such as IT, business, management, and finance, have different levels of AI chat use.

Conclusions

The study aimed to test students' acceptance of ChatGPT using the UTAUT2 framework, which is extensively used in technology acceptance research. The study confirmed the significant impact of “Habit”, “Performance Expectancy”, and “Hedonic Motivation” on “Behavioral Intention” to use ChatGPT by students. Since ChatGPT use in higher education is not yet explored, there are many recommendations for future studies. For example, future studies can test the scale used in this study and further develop it in the future.