Keywords

Introduction

Generative Artificial Intelligence (GAI) was not a popular discussion topic among faculty in Higher Education (HE) until the emergence of tools such as ChatGPT. Since 2021, the academic discourse both in relation to policies in the use of AI, but also in relation to the potential opportunities of AI for education, has started to rise as a research topic (Southworth et al., 2023). Given the widespread availability of generative AI tools like ChatGPT, it is imperative for Higher Education Institutions (HEI) to carefully examine the practical applications and possible difficulties of AI for both professors and students. Although we must acknowledge the potential applications of AI, it is crucial to address the issue of regulating its use in the context of university work undertaken by undergraduate and graduate students. Rudolph et al. (2023) point to the challenges facing HE institutions that continue to use traditional assessment strategies when the availability of tools such as ChatGPT makes it difficult to evaluate the originality of student work. Conversely, AI can be used to scaffold student learning and create more personalised HE experiences.

GAI typically refers to advanced technology that integrates deep learning models, trained on extensive datasets, gathered from various public sources, user-generated content, licensed third-party data, and information created by human reviewers (OpenAI, 2023). This technology processes human inputs, commonly known as prompts, and generates outputs that closely mimic human-generated content, predominantly in the form of text and images (Lim et al., 2023). Due to their large scale, software developers building these tools utilise models that frequently lack direct insights about the quality and type of data used for training. Additionally, they are often unable to meet data retention or privacy requirements given their inability to store these data models in their independent computing environments. Hence, similar to using a search engine, a typical user currently relies on remote server interactions when exchanging data through AGI tools. This raises significant concerns regarding privacy and the potential for information leakage. This issue is particularly acute with AGI tools like ChatGPT, which require more detailed text input, in contrast to typical search engines that respond to relatively brief search queries.

In this chapter, we address these different domains by analysing the uses of AI in HE, with a special focus on generative AI. The second section addresses AI for language learning and translation in HE. The third section explores the use of conversational agents (i.e., chatbots) at the university level, while the last section addresses the responsible uses of AI in HE with a special focus on the role of assessments.

Uses of AI in Higher Education

In their 2023 literature review, Baidoo-Anu and Ansah focused on General Artificial Intelligence (GAI) and education-related papers published in English-language, peer-reviewed journals. The review aimed to achieve two primary objectives: to evaluate the various methods of interacting with ChatGPT, and to discern the advantages and disadvantages of integrating GAI into educational practices. Within this framework, the authors introduced a typology categorising the observed benefits of employing GAI in education, encompassing areas such as personalised tutoring, Automated Steady Grading (AGS), language translation, interactive learning, and adaptive learning. Simultaneously, the review outlined a series of limitations associated with these AI applications, including the lack of human interaction, limited understanding of the technology, potential biases in training data sets, lack of creativity, dependency on the data available or generated for AI training, lack of contextual understanding, limited ability to personalise instruction, and privacy concerns. This detailed exploration of both benefits and limitations contributes to a more nuanced understanding of the impact of General Artificial Intelligence (GAI) in the realm of HE andragogy.

Several exploratory studies have applied prompt engineering (Lee et al., 2023) to more effectively examine how the GPT-3.5 model aligns with educational objectives and its suitability for such purposes. The primary methodology involves analysing the model's responses and evaluating their congruence with educational objectives. This assessment is conducted through a self-study approach as introduced by Hamilton et al. (2009). Cooper (2023) utilised ChatGPT by presenting it with a series of questions designed to elicit responses providing practical guidance for teachers on classroom applications. Findings suggest that ChatGPT performs well in generating teaching units, rubrics, and quizzes. Additionally, the results indicate that ChatGPT can aid educators in the creation of science education units structured around the 5Es model (Engage, Explore, Explain, Elaborate, Evaluate), thereby assisting in the transition from initial ideas to fully developed educational units.

Similarly, Qadir (2023) examined ChatGPT's role in engineering education, underscoring its potential as a generative AI tool across technically demanding educational settings. The research employs structured prompts as a method for eliciting detailed AI responses that facilitate various real-life educational applications. Although not based on a specific pedagogical framework, the study provides empirical evidence of ChatGPT's utility in engineering education. Its applications span technical subjects like coding and mathematics, creative writing, virtual tutoring, personalised learning, test preparation, and language learning. Conversely, some of the disadvantages include the lack of human interaction in providing personalised feedback to learners. Moreover, as the study involved an older model of ChatGPT (i.e., GPT-3.5 architecture), concerns regarding reliability, plagiarism, and hallucinative misinformation were noted as potential shortcomings when using automated feedback. The paper also emphasises the growing importance of prior knowledge and critical thinking, given their importance in creating prompts that generate quality responses.

Chan (2023) has studied the use of text-generative AI technologies in Hong Kong universities to develop a framework for the integration of AI into education. The study engaged 457 students and 180 faculty and staff members. Chan has identified three dimensions: the pedagogical dimension, the governance dimension, and the operational dimension in the integration of AI in higher education. Within the context of HE andragogy, Chan’s framework (2023) highlights the importance of reengineering the assessment process given the innovative methodologies made possible through the use of AI. In this sense, both the automatic analysis, as well as the exploitation of learning analytics for assessment purposes (Ouyang et al., 2023) were examined. Similarly, the pedagogical dimension is also where Chan stresses the importance of developing the transversal competencies considered critical for future success in the innovation economy (Septiani et al., 2023).

Within the governance dimension, senior leadership plays a pivotal role in addressing the complex considerations related to AI implementation in education. This encompasses strategies to understand, identify, and prevent academic misconduct and ethical dilemmas facilitated by AI, as well as the establishment of robust policies and protocols for data privacy, transparency, accountability, and security in AI usage. In this sense, the AI Act, developed at the European level, aims to develop trustworthy AI (Laux et al., 2023). The AI Act postulates different requirements for ‘trustworthy’ AI: human agency and oversight, technical robustness and safety, privacy and data governance, transparency, diversity, accountability, non-discrimination and fairness, societal and environmental well-being, and accountability (EC, 2019). Furthermore, the governance dimension extends to AI attribution and the clear definition of roles and responsibilities for the technology’s implementation and management, thereby ensuring accountability for its ethical use within the institution. Equally significant is the commitment to ensuring equity in access to AI technologies, achieved through implementing measures that guarantee fair and inclusive access to AI resources for all students and faculty, while also addressing potential disparities in AI utilisation across different demographic groups.

The operational dimension, involving teaching and learning stakeholders as well as IT staff, centres on the practical aspects of AI implementation in university settings. This includes establishing robust monitoring mechanisms to assess the effectiveness of AI integration and continuously evaluating its impact on teaching, learning, and overall educational outcomes. Moreover, this dimension encompasses comprehensive training programmes and ongoing support structures that enhance AI literacy among faculty, staff, and students, while addressing the challenges and benefits of proficient AI use across the university.

Sabzalieva and Valentini’s (2023) guide on the use of generative AI in HE develops an introduction to the technology and provides different use cases as an interactive tool fulfilling the role of tutor, socratic opponent, or even co-designer.

AI for Language Learning and Translation

The prevalent use of English in research studies, publications, and competitive grants at the international level creates an inclusivity barrier, not only for faculty and students, but also for administrative staff who lack the necessary proficiency in English to fully participate in the academic process (Ingvarsdóttir & Arnbjörnsdóttir, 2013). This linguistic dominance reduces the opportunities and competitiveness of non-native English speakers and limits access to learning resources for students who lack English fluency. As such, AI translation tools, such as Grammarly or Quilbot, can play a pivotal role in facilitating accessibility, but also improving the overall quality of written work, from both a grammatical and contextual perspective. Implementing AI-driven translation tools can reduce language barriers such as in the case of the MSc Smart EdTech programme at the Université Côte d’Azur, where most of the students and faculty are using English as a second language. With an international cohort representing eighteen different countries, the use of real-time translation tools during synchronous, online activities helped create a more inclusive learning environment for non-fluent, English speakers. Similarly, the use of automatic note taking tools has facilitated the creation of meeting minutes and made documenting projects easier in the context of research initiatives such as the Horizon AugMentor project.

Chatbots in Higher Education

Expanding from the pioneering work of Eliza (Weizenbaum, 1966), acknowledged as the first chatbot in academic research, the exploration of chatbot-assisted learning environments has progressed significantly over the last few decades. Nevertheless, it is in recent years that chatbots have experienced a substantial surge in both usage scenarios and research within the educational realm, reaching a pinnacle in 2023 (Hwang & Chang, 2021). The most common application of chatbots in educational settings is as a tool used to interact with predefined content learning paths, often referred to as guided learning (Akcora et al., 2018). The current availability of generative AI chatbot services and chatbot application programming interfaces (API) enables educators to add a new layer of chatbot capabilities that can support active learning approaches in education (Lo., 2023). Chatbots were created to enable natural language interaction between humans and computers. As such, chatbots can be considered as computer programmes that aim to mimic some aspects of human interaction supported by machine conversation systems, virtual agents, dialogue systems, and personal assistants all with the goal of supporting the end-user (Suhaili et al., 2021). The systematic literature review by Okonkwo and Ade-Ibijola (2021) outlines the diverse applications of chatbots in HE, including teaching and learning (66%), research and development (19%), assessment (6%), administration (5%), and advisory (4%).

Integrating chatbots into HE settings enables universities to address the specific uses outlined in the UNESCO table on the use of generative AI (see Table 10.1). This is possible due to the versatile nature of these tools and the adaptive experience they offer to users. Furthermore, the ability of chatbots to understand and generate human-like responses, known as Natural Language Processing (NLP), creates a more intuitive way for students and educators to interact with the tools (Maher et al., 2020; Rath et al., 2023), while also facilitating the personalisation of the learning experience based on individual student needs (Younis et al., 2023).

Table 10.1 Use of generative AI in higher education according to UNESCO

The inclusion of NLP capabilities in the training of chatbots has benefited developers and pedagogical staff seeking to personalise their use for HE environments. Additional refinements to the personalisation process occur during tokenisation,Footnote 1 where text is converted into smaller units in order to make the chatbots more efficient and effective (Bhartiya et al., 2019). Once trained and generating responses with a high level of reliability, developers continue to monitor the chatbot’s performance and provide iterative training to ensure that it is functioning smoothly and providing accurate information. Moreover, Tsivitanidou and Ioannou (2021) considers the potential of chatbots to support certain types of learning scenarios in HE.

In their guidelines, Chocarro et al. (2023) explain the desirability of empowering teaching and administrative staff to effectively use AI, even as its use continues to face questions regarding user digital literacy, ethics, data privacy, and how these tools impact current pedagogical strategies.

It is essential that a comprehensive artificial intelligence training program for university students and staff includes learning skills related to using, developing, and implementing chatbots. This inclusion is crucial at various levels:

  • Problem Identification: Universities should support the pedagogical empowerment and the AI acculturation of operational staff to identify opportunities and problems that can be facilitated using chatbots.

  • Theoretical/Practical Framework: Universities should provide assistance in the creation and development of chatbot-based assisted learning scenarios based on the needs of educators and learners.

  • Ubiquitousness: Universities should ensure that the use of chatbots is democratised, widespread, and reflective of the latest technological and training iterations.

  • Practical use: Universities should assist students in learning the skills necessary to seamlessly integrate chatbots into their learning environment.

  • Assessment: Universities should continuously assess and refine their chatbot training models following a holistic set of quality standards.

When contemplating the implications associated with the development and application of chatbots in educational settings, educators have the responsibility to consider the role biases play in their usage and development. For instance, tokenisation, a key aspect in data processing, requires careful attention to mitigate instances of information biases and to ensure that the chatbots maintain a pedagogical perspective. Ensuring that the tokenisation of data is executed without bias is imperative, as it directly impacts the effectiveness, accuracy, fairness, and equity of the educational experience as facilitated by the chatbot (Akter et al., 2021). Furthermore, educators must advocate for the availability of chatbot APIs capable of seamlessly functioning in different languages, thereby fostering inclusivity and accommodating linguistic diversity (Mogavi et al., 2023).

Finally, the shift in AI chatbots development APIs to no-code and low-code, especially OpenAI's GPTs in November 2022 (Lim et al., 2023), marks a significant milestone in democratising chatbot development for education. This shift has a notable impact on the development and use of chatbots as it allows individuals to create their own customised AI tools without having extensive knowledge of software development or programming. In the context of educational use cases, this removes many of the significant barriers preventing educators from leveraging their unique expertise and training in the creation of their own AI tools.

Responsible Use of Generative AI Tools in Academia

In November 2022, OpenAI made ChatGPT publicly available for free, employing the highly advanced GPT-3 model as its backbone Large Language model (LLM). By January 2023, OpenAI announced that ChatGPT had accumulated over 100 million users, setting a new global record as the fastest-growing application to date (Lim et al., 2023). The exponential growth of ChatGPT had a significant impact on the education sector, as students and educators around the world began exploring the app's novel functionalities. Delivered through an intuitive and user-friendly chatbot interface, ChatGPT's text translation, question-answering (Q&A), and text generation capabilities introduced new opportunities and challenges to modern learning and teaching values, norms, and methodologies. This development has been met with mixed reactions from the educational community, prompting educational institutions worldwide to establish ad hoc committees of experts tasked with revising their ethical frameworks, guidelines, and recommendations concerning the use of Generative AI (GAI) in education and pedagogy.

Russell Group universities provide a comprehensive framework dedicated to promoting the ethical and responsible utilisation of GAI tools within academic settings. These institutions are steadfast in their commitment to following established guidelines for the ethical use of AI tools in education, as outlined by the principles set forth by the Russell Group. This commitment involves fostering AI literacy among both students and staff and empowering educators as they guide students in the effective and responsible use of generative AI tools. Additionally, the Russell Group universities actively engage in reviewing and adapting curriculum, teaching methods, and assessment practices to seamlessly integrate the ethical use of generative AI and ensure equitable access for all. This commitment extends to upholding academic rigour and integrity, while also fostering collaboration with other institutions to share best practices in response to the evolving technological landscape and its educational applications.