1 Introduction

Education is a main factor in the development, economic advance and welfare of communities [20]. Early stages of K-12 education are particularly important in acquiring the learning skills that will last with the students for the rest of their lives [15]. Bloom’s one-to-one tutoring and learning for mastery theories [8, 9] have always been respected by the educational community. Yet, the impracticality of their application, in conventional classroom teaching and teachers-to-students ratio, has always been a limitation [9]. It was not until the era of Information Communication Technology (ICT) and Artificial Intelligence (AI) that these theories became promisingly applicable [10]. There is a growing trend towards strengthening the dependence on distant-learning; expanding and developing E-learning platforms and; most importantly, introducing adaptive personalized features based on the needs and learning pace of each student [48].

A wide variety of E-learning and tutoring platforms provide the same content to all learners [34]. Fewer platforms provide personalized recommendation and adaptation features [48]. Learning styles can be defined as a set of cognitive, emotional, characteristic and physiological factors that serve as relatively stable indicators of how a learner perceives, interacts with, and responds to the learning environment [30]. Cognitive level stands for the learner’s level of knowledge. Adaptive personalized e-learning research focused on adult learners and considered learning styles and cognitive level as adaptation means [7, 42, 50, 62]. Most studies considered only one adaptation means and a corresponding target with a rule-based adaptation strategy [62]. As the number of adaptation means and targets increase, AI is needed to mimic human decision making for such various situations [61].

Many studies recommended multisensory learning [13] and offering students a set of multimedia learning material to choose from [32]. An adaptive personalized learning environment, which is designed to regulate different modalities based on their cognitive/learning styles, would be even better [32]. For students’ learning styles, various models and questionnaires exist but they were designed and validated for adults [28]. Hence, their modification to fit younger people represents a challenge. Since VARK provides a questionnaire specifically designed and validated for young children [35, 60], it can be used to initiate this dimension of our learner model. Based on the multiple intelligences and sensory modalities, it is the most convenient questionnaire to describe children learning behaviors as percentages of the different modalities [40, 49, 58]. In addition, it is one of the most effective learning styles dimensions utilized in personalization and adaptation [1].

On the other hand, gamification: “the use of game design elements in a non-game context”, is now the new trend in learning even in traditional classroom context [17, 29, 41, 59]. In games, children are more eager to accept challenges, rules, tasks, and feedback. They are even eager to enhance their skills to open higher levels of the game. Such game elements foster the creation of an immersive experience leading to higher involvement and engagement. Digital Game-Based Learning (DGBL) in K-12 education highly influences student’s motivation, engagement, and learning outcomes [17, 29, 41]. In a case study of elementary school students, teachers experienced a 13% increase in students’ attendance in mathematics classes after implementing gamification techniques inside the traditional classroom [2].

For students’ knowledge level, Bloom’s taxonomy classifies educational learning objectives into levels of complexity and specificity: Remember, Understand, Apply, Analyze, Evaluate, and Create [33]. It has been the primary focus of most traditional education and is frequently used to structure curriculum learning objectives, assessments and activities. Exercises difficulty scaffolding was implemented in several e-learning research studies based on cognitive level [43]. Yet, few studies provided a clear objective for this adaptation and they used rule-based adaptation [43]. Automatic detection of the suitable exercises difficulty enhances engagement and learning efficiency compared to self-regulated methods [45].

AI and machine learning have found their way to many applications such as: Internet of Things (IoT) [65], Intelligent Transportation Systems (ITS) [16], solution of mathematics problems [3, 4] and education [48]. RL is a form of machine learning that maps situations to actions in order to maximize the long-term reward. It differs from supervised and unsupervised machine learning approaches in the presence of learning agents. The learning agent senses the environment (despite its uncertainty), chooses the action that will maximize the rewarding function and updates the state accordingly [55]. Consequently, RL represents a very appropriate setup for an adaptive personalized e-learning environment and the proposed platform. DQN-RL leverages advances in deep learning to learn policies from high dimensional sensory inputs. Specifically, it learns with raw data using convolutional neural networks, instead of low-dimensional features vectors.

This paper proposes an Adaptive Personalized Platform for an Effective and Advanced Learning (APPEAL) for school students. APPEAL makes use of the previous literature from different aspects and aspires to bridge the gaps between different disciplines and overcome their limitations. It utilizes a novel combination of adaptation means, targets, strategies and goals. As adaptation means for learner modeling, APPEAL includes cognitive (learning style); behavior (effort, support and performance) and affective (engagement) categories of the student personal traits. A dynamic learner model, which detects and corrects the learner model in terms of learning style and cognitive level, is developed. For personalization and adaptation, APPEAL utilizes a DQN RL-based AI implementation and an online rule-based decision making based on student interactions as the adaptation strategy. It aims at achieving a personalized learning experience covering two main adaptation targets. First, it provides adaptive presentation through VARK for one group and gamification for the other group. Second, it provides adaptive exercises navigation and feedback. Adaptive navigation is achieved through scaffolding of exercises difficulty in Bloom’s taxonomy by skipping/hiding/reattempting. In addition, adaptive feedback is achieved through hints, attempts, and feedback messages. APPEAL aims at achieving the following adaptation goals: more effective learning with reduced time spent and improved grades and satisfaction level. To the best of our knowledge, no previous research study compared the learning outcome of using a combination of VARK modalities and gamification. Besides presenting the novel platform, APPEAL, with adaptive personalized presentation, navigation and feedback and its simulation based validation, the following research questions are assessed for both VARK and gamification groups in a pilot experiment:

  • RQ1: Does APPEAL help improving aggregated-level academic performance and learning effectiveness indicators (pre- and post-test scores, completion time and learning efficiency) and how much improvement is achieved?

  • RQ2: Does APPEAL help improving the data dispersion for these academic performance and learning effectiveness indicators and how much improvement is achieved?

  • RQ3: Does APPEAL help improving disaggregated-level academic performance and learning effectiveness indicators for each student on lesson and exercises level?

  • RQ4: Does APPEAL achieve good student engagement and satisfaction indicators?

The rest of the paper is organized as follows. Section 2 reviews some related researches on adaptive and personalized e-learning including e-learning systems for school students, gamified content and RL-based implementations. Section 3 demonstrates the adaptive personalized user interface provided by the front-end of APPEAL for both VARK and gamification groups displaying several screenshots. Section 4 is more concerned with the back-end architecture and adaptation algorithms: adaptive personalized RL-based VARK presentation and online rule-based exercises navigation. Section 5 presents and discusses the learning effectiveness results for the two experimental groups (VARK and gamification) and provides answers to the research questions of Section 1. Section 6 concludes the work and suggests future research directions. Finally, a list of all used symbols is given in the Appendix.

2 Literature review

In many countries, ICT application is still less common in K-12 education compared to college and adult education [66]. Focusing on student modeling and corresponding adaptation for school students, several relevant studies in the last decade were reviewed. Table 1 provides a summary of these papers focusing on the following dimensions: purpose, design/methodology/approach and findings/implications/ recommendations. Although [21, 63] do not focus on dynamically detecting nor correcting the learner model; they show how effective the teaching style is in the academic performance of the learners for school students. Study [47] cluster/detect learner model in order to assess the preferable adaptation strategies without implementing any adaptation. Studies [39, 56] showed the importance of learning style correction through a dynamic learner model; however, the corresponding adaptation strategies were not investigated. Studies [11, 18] focus on the content adaptation based on static learner model for cognitive level detection. Although [14] considered dynamic changes of the model with the interactions, this system did not consider the learning style. Study [12] proposed a model-tracing intelligent tutor that interpret students’ mathematical problem-solving behaviors in fractions topic and provide adaptive personalized lower level exercises and feedback. Study [25] provided adaptive exercises navigation (skipping) and feedback (hints), which resulted in scores improvement from pre- to post-test. Study [24] discussed adaptive e-learning focusing only on exercises omission for high performing students. Yet, the system [24] did not adapt the learning content according to any learning style model. Other studies such as [19] used a unified presentation, as opposed to personalized, but in the form of mathematics worked example videos. Yet, it led to better results than the traditional method and a students’ survey revealed their satisfaction.

Table 1 Summary of the reviewed papers targeted at school students

In summary, the reviewed relevant works focusing on school students suffer from the following general limitations:

  • Most of them were based on static learner models through questionnaires and even did not proceed to implement an adaptation module or provide adaptation rules.

  • Most of them did not cover all elements of the learning process or branches of adaptation of content presentation and exercises navigation.

  • Some of them did not focus on the implementation technique of the learner or adaptation models and most of them did not employ AI-based solutions.

  • Most of them did not provide clear evidence or enough performance metrics of the learning effectiveness of their proposed systems.

  • Most of them did not consider the affective state of the student, engagement or satisfaction surveys.

As for gamification, interactive games yielded an improvement in post-test performance in English teaching in [44], either depending on computer mouse or motion sensing. After-school remedial instruction of mathematics was considered in [37], where the game-based group outperformed the video grou.p although both resulted in good performance. The effect of gender on the verbal and nonverbal engagement behavior in game-based learning was explored in [22], which observed that male students were more engaged.

From the viewpoint of adaptive and personalized e-learning implementation, RL is proven to be an effective approach [5, 6, 26, 27, 36, 51, 52, 57]. Yet, these works also suffer from the following limitations:

  • Objectives of the studies focused on either learning style detection or academic level improvement. However, there is a lack of studies that consider both dimensions.

  • On the level of academic performance improvement, the reward function, which directs RL towards the most suitable action, usually focused on either improving the students’ performance (grades) or reducing their effort (time on task) but not both.

  • These studies mostly targeted college level and adult learners. RL was utilized for elementary school students’ learning assistance in computer skills in [23]. Yet, the only provided adaptation target was adaptive personalized feedback with the goal of saving teachers’ time.

  • Deep RL has not yet been widely employed in this field although it shows promising results [53]. The authors of [53] aimed at improving students’ learning performance and reducing their time on task, but not simultaneously. In addition, the adaptation was limited to deciding between worked examples or problem solving.

APPEAL attempts to overcome these limitations by developing an adaptive personalized e-learning platform targeted at school students with a DQN-RL-based implementation. A dynamic learner model is developed and, accordingly, content presentation, exercises navigation and feedback are adapted altogether. APPEAL’s goal is to improve the students’ learning performance, reduce their time on task and achieve a high satisfaction level. To asses the achievement of this goal, various aggregated-level, statistical and disaggregated-level academic performance, learning effectiveness and satisfaction level indicators are investigated. To the best of our knowledge, no previous research study compared the learning outcome of using a combination of VARK modalities and gamification.

3 APPEAL user interface

Figure 1 shows the general architecture of APPEAL, which is built as a dedicated website based on Moodle. As an adaptive personalized e-learning platform, APPEAL consists of the front-end, back-end, and the Application Programming Interface (API) between them. The front-end includes all user interface and it is where all student interactions take place. The back-end includes all the rule-based and machine intelligence-based algorithms that make all the decisions, i.e., what to present and when for each student based on his/her interactions. The API is the interface through which the first two components: front-end and back-end communicate. Details of the three components of APPEAL are given in this section and the next section.

Fig. 1
figure 1

APPEAL general architecture

The student’s journey through APPEAL is shown in Fig. 2. It starts by logging into his/her account. The student takes an initial VARK questionnaire and proceeds to the lesson presentation based on his/her group (VARK versus gamification). For each unit of the curriculum, the student takes a pre-test to determine his/her entry level. The lessons are presented into different presentation styles/modalities or gamified including explanation and examples. After each lesson, there is a set of exercises that cover Bloom’s taxonomy 6 categories (T1 to T6). Each taxonomy category is further measured by three levels of exercise difficulty: Easy-Medium-Hard (E-M-H). Only a subset of each taxonomy’s exercises is displayed to the student based on his/her level. Both the easy and medium levels with lighter colors are subject to hiding/skipping if the student interactions reflect an advanced level. Reattempting some levels with equivalent exercises from the question bank can also take place when needed till mastery. The aim is to improve student level such that he/she becomes capable of solving hard exercises despite his/her entry level. Throughout these steps, the student interactions: time spent, attempts, hints and grades are recorded. In addition, they are communicated to the back-end adaptation algorithm to modify its decisions accordingly. After each unit, the student takes a post-test to determine his/her exit level. Finally, students are asked to fill out a satisfaction-feedback survey.

Fig. 2
figure 2

Flowchart of the student’s journey through APPEAL

The front-end of the official website APPEAL is designed based on Moodle. The content and tests of Grade 3 mathematics was prepared under the supervision of school teachers. All lessons were prepared in the four presentation modes/styles: visual, aural, text and kinesthetic. Figure 3 shows samples from the four different lesson presentation modalities including explanation and examples for VARK group. The back-end algorithms (Section 4) determine which presentation will be given to the student in an adaptive manner.

Fig. 3
figure 3

Four different lesson presentation modalities for VARK group

The gamification part of APPEAL focuses on the student affective state by targeting increased engagement. The design includes features like offering virtual markets, a pointing system, progress bar, and interactive quizzes in the form of challenges and competitions to enhance students’ motivation, eagerness to learn more and engagement. The lessons are prepared as Shareable Content Object Reference Model (SCORM) e-learning software products. Figure 4 shows the gamification interface including home screen with start, help and exit icons as well as the screens that appear upon clicking on these buttons. Figure 5 provides an example on the details of units and lessons navigation in a storytelling form, where the student is the main character navigating from an island to another. The function of each icon and the features provided by them are also shown including music, controls, replay, navigation, … etc. In addition, locked levels encourage the students to proceed in the learning process to unlock them and feedback is provided in the form of score follow-up in the form of a pointing system with progress bar. Figure 6 demonstrates the main features of explanation and examples with snapshots from the gamified scenario of a lesson.

Fig. 4
figure 4

Home, start and help screens for gamification group

Fig. 5
figure 5

Units and lessons navigation and features for gamification group

Fig. 6
figure 6

Lesson (a) explanation and (b) examples samples for gamification group

Exercises were provided and classified by the teachers based on Bloom’s taxonomy categories (T1 to T6) according to their objectives and (E-M-H) according to their difficulty level. The structure of the exercises bank per lesson is shown in Fig. 7, where 3 × 6 = 18 exercises cover the three difficulties for each taxonomy level. Teacher accounts can add content and exercises and access the student data of their classes. They add the exercises to the corresponding classification based on the nature of questions under each Bloom’s taxonomy level. The academic level of the learner among E, M and H is initialized by his/her level in the pre-test and updated from previous interactions. In addition, the back-end algorithms (Section 4) determine which difficulty level he/she will start at or navigate to. Figure 8 shows some screenshots from the exercises and the feedback given to the students at different situations, which are common for both groups.

Fig. 7
figure 7

Structure of the exercises bank per lesson

Fig. 8
figure 8

Sample screenshots from the (a) exercises and (b) feedback to the students in both groups

4 Adaptation algorithms

The back-end of APPEAL receives the student state based on his/her interactions with the front-end, as well as the student tracking history, as shown in Fig. 9. Then, it adapts content presentation according to the student’s learning style using an RL algorithm. In addition, it adapts the exercises navigation based on his/her knowledge/academic level using an online rule-based decision making algorithm. We start by explaining the exercises navigation as the student state is mostly obtained from his/her performance in the exercises.

Fig. 9
figure 9

APPEAL back-end architecture and communication with front-end

4.1 Adaptive personalized exercises navigation and feedback

A unified online rule-based decision making algorithm based on student interactions is used. Depending on the student behavior: effort (attempts), support (hints) and performance (grades) as adaptation means, adaptive personalized exercises navigation aims at determining the appropriate exercises difficulty within each Bloom’s taxonomy level. The algorithm determines the difficulty level of the exercise that appear to the student at entry level per taxonomy category (tax), where difficulty ∈{1,2,3} corresponding to E,M,H, respectively. For simplicity, the difficulty is initialized by the pre-test score of each unit for all taxonomy categories as follows:

$$ {difficulty}_{Pre}= \begin{cases} 1, \quad & \qquad {Pre} < 50, \\ 2, \quad & \qquad 50 \leq {Pre} <75, \\ 3, \quad & \qquad 75 \leq {Pre}. \end{cases} $$
(1)

Throughout the student’s journey through APPEAL, the following algorithm keeps track of his/her level per taxonomy category based on a penalized grade, gradepenalized, and decides the difficulty of the exercise to be displayed next. The decision or action corresponds to improvement/upgrading (+ ), no change (0), or deterioration/downgrading (−). The penalized grade implicitly includes the attempts and hints. gradepenalized= 1 when the student gets the correct answer(s) by himself/herself. With every hint up to the available 4 attempts, gradepenalized is computed from grade by subtracting 0.2, where grade ∈ [0,1]. If the last attempt failed, the answer is displayed and gradepenalized= 0. Two thresholds are defined as thg1= 0.4 and thg2= 0.8. Separate decisions are made for each taxonomy category as given in Algorithm 1.

Students with penalized grade less than 0.4 have not provided the correct answers and are downgraded to a lower level (− 1). Students with penalized grade between 0.4 (inclusive) and 0.8 (exclusive) have reached the correct answers by themselves after given two or three hints and can be upgraded to one level higher (+ 1). Students with penalized grade greater than or equal to 0.8 have reached the correct answers directly or after a single hint and deserve to be upgraded to two levels higher (+ 2). Correct answers of the hard exercises in all taxonomy categories are considered as a pre-requisite for the next lesson. Upon reaching the hard exercises in all taxonomies, the student must answer them to unlock the next lesson. If all hints were used and the correct answer is displayed to the student, unlocking the next lesson requires answering an exercise of the same level. This guarantees the aim of APPEAL that the student level improves to the level that corresponds to hard exercises in all six taxonomy categories. Then, next lesson difficulty is initialized from pre-test score, and so on.

figure a

4.2 Adaptive personalized VARK presentation using RL

Adaptive personalized presentation can be formulated as a stochastic control problem, more specifically Markov Decision Process (MDP), where the state of environment is represented by the interaction of the student with the E-learning platform. Such interactions are partially random and partially affected by the platform. They have stochastic features especially at an early age where the personality traits and character are still being developed. The MDP problem is optimally solved using dynamic programming and practically solved using function approximation or deep reinforcement learning due to the curse of modeling and curse of dimensionality issues [55]. MDP is defined by a tuple < S, A, R, T, γ >, which consists of a set of states S, a set of actions A, a reward function R(s, a), a transition function \(T (s,a,s^{\prime })\) that equals the probability distribution \(P(s^{\prime }|s,a)\), and a discount factor γ. In each state sS, the agent takes an action aA. Upon taking this action, the agent receives a reward R(s, a) and reaches a new state \(s^{\prime }\), determined from \(P(s^{\prime }|s,a)\).

A policy π specifies for each state which action the agent will take. The goal of the agent is to find the policy π, which maps states to actions, that maximizes the expected discounted total reward over the agent’s lifetime. The value Qπ(ϕ(s),a) of a given state-action pair (s, a) is an estimate of the expected future reward that can be obtained from (s, a) when following policy π, where ϕ(s) = s in our case. The optimal value function Q(s, a) provides maximal values in all states and is determined by solving the Bellman equation:

$$ Q^{*}(s,a)=E\left[R(s,a)+\gamma \sum\limits_{s^{\prime}}P(s^{\prime}|s,a) {\max}_{a^{\prime}} Q^{*}(s^{\prime},a^{\prime}) \right]. $$
(2)

The optimal policy π is then π(s) = argmaxaAQ(s, a).

4.2.1 State-action-reward < s, a, r > triplet

APPEAL provides personalized presentation using an RL approach based on both cognitive (learning style) and behavior (effort, support and performance) categories of the student personal traits. As previously explained, the learning agent of RL senses the current state from the environment, chooses the action that will maximize the rewarding function and updates the next state accordingly [55].

  • State is defined by the student interactions with the domain module, i.e., time spent per lesson (effort), scores per exercise (performance), number of hints checked (support), and number of attempts of the same taxonomy category and difficulty level (effort). Student state definition is also given in Figs. 9 and 10.

    Fig. 10
    figure 10

    RL agent

  • Action is defined by the adaptation module as the detected learning style (VARK) with the corresponding presentation to the lesson explanation material in the form of four binary digits, where only styles with “1” are presented while others with “0” are hidden.

  • Reward is defined as the improvement in the academic performance of the student and the saving in the time spent, hints checked and number of attempts per lesson, which are supplied as the input state from the API.

4.2.2 RL agent

The utilized algorithm is DQN with experience replay given by Algorithm 2 [38]. DQN approximates the value function Q(s, a) with a deep neural network that outputs a set of action values Q(s, a;𝜃) for a given state input s and action a, where 𝜃 are the parameters of the network. There are two key components of DQN that make this algorithm work. First, it uses a separate target network \(\hat {Q}(s,\cdot ;{\theta }^{-})\) with separate weights 𝜃, which are copied every C steps from the regular network, so that the target Q-values are more stable. Second, the agent adds all of its experiences to a replay buffer D, which is then sampled uniformly to perform updates on the network. An experience is the tuple \(<s,a,r,s^{\prime }>\).

An 𝜖-greedy policy is used to generate actions as follows. Select the action that gives the maximum Q-value with probability (1 − 𝜖) or select a random action with probability 𝜖. The probability 𝜖 decreases through learning as the number of iterations increases, which is known as annealing. All the interactions of the user/student are stored in memory. The next action is determined by the maximum output of the Q-network a = argmaxaAQ(s, a, 𝜃) when using the optimal policy. The loss function is defined as follows after sampling a random minibatch of transitions \((s_{j},a_{j},r_{j},{s_{j}}^{\prime })\) from D:

$$ y_{j}=\left\{\begin{array}{ll} r_{j} & \quad \text{if episode terminates at step } j+1, \\ r_{j}+\gamma {\max}_{a^{\prime}} \hat{Q}({s_{j}}^{\prime},a^{\prime},{\theta}^{-}) & \quad \text{otherwise}, \end{array} \right. $$
(3a)
$$ \text{loss} = \left( y_{j}-Q(s_{j},a_{j},\theta)\right)^{2}. $$
(3b)

The minimization of the loss function is performed using gradient descent.

figure b

The implemented RL agent shown in Fig. 10 has two main modules: a feature extraction module and action decoder module. The feature extraction module consists of a series of fully-connected layers (dense layers). To improve the stability of training and speed of convergence, each fully-connected layer is followed by a batch normalization layer. This layer lets the network decide the best normalization strategy for the output of the previous layer before going through a ReLU non-linearity.

4.3 Simulation-based validation

The deep network was implemented using PyTorch 0.4 framework under Python 3.6 and trained using NVIDIA GTX1070TI graphics card for one million episodes. A four-layer neural network was used with [128, 128, 64, 32] neurons per each layer, respectively. A batch size of 4096 state-action-reward-next state quadruples, a learning rate of 7e-4, an experience replay memory of one million quadruple, and a discount factor γ of 0.999 were used. An 𝜖-greedy policy was applied and starts with a high 𝜖 of 0.9 to encourage state exploration. Then, it decays exponentially as the number of training iterations increases until it reaches 0.005, which exploits the best action taken by RL agent.

Before testing the model on the actual school students, it was trained using a validated student simulator. The action of the RL agent and the output of Algorithm 1 are passed to the simulator at the beginning of each episode. As the simulator outputs, the state, i.e, interactions measurements analogous to those expected by real students, are produced. A corresponding reward value is also produced to aid the convergence of the RL algorithm. Both the simulated state and reward are set based on previous experience of the teachers with students in classroom and non-adaptive e-learning. Initial learning style and initial level (E, M or H) per taxonomy are considered ground truths about the student once generated. While these parameters are randomly generated in the training phase using the simulator, they are determined from the VARK questionnaire and pre-test for real students. The rest of the configuration parameters include thresholds for the components of the student state definition (time, scores, hints, attempts) and level development (academic level improvement, i.e., scores improvement in addition to time, hints and attempts savings) corresponding to the reward as explained in Section 4.2.1.

Each episode starts with a simulated student selected randomly. Then, it is terminated in two ways: either by reaching the final state in which the student masters the lesson and is ready to unlock the next lesson or the depth of unrolling the episodic tree has reached one hundred states (iterations). The decision time step is the completion of at least one exercise for each of the remaining taxonomies, since the student may have completed the advanced level (hard) exercise in some of the taxonomies. Every step or iteration, a new state-action-reward triplet takes place.

The DQN convergence was validated, where the reward is maximized and the episode is faster for high performing students. The algorithm was validated to act correctly for several scenarios, where the student starts from an entry level corresponding to his/her pre-test score and is downgraded/upgraded based on the obtained penalized grades matching his/her initial ground truth level.

5 Pilot experiment and experimental work

This section summarizes some state-of-the-art metrics to evaluate learning effectiveness and their results for the two experimental groups (VARK and gamification) and provides answers to the research questions of Section 1. COVID-19 pandemic and social distancing procedures did not allow comparing against control group (traditional classroom teaching only). The number of students who completed the whole experiment is 26 students in grade 3, whose age ranges from 9-10 years old . Although the sample size is small, some of the reviewed studies reported their results on case studies of the same range [21, 22, 36, 57]. The students were given IDs from 1 to 13 in VARK group and 14 to 26 in gamification group, where the subdivision among the two groups is random. The experiment is carried out for two units covering 8 lessons over the course of around 2 months. Unit 1 is entitled Multiplication and Division with the lessons: Multiplying by 10, Multiplying by 100, Multiplying by 1000, Multiplying a 2- digit number or more by a 1-digit number, Even numbers and odd numbers, and Dividing a number by a 1-digit number. Unit 2 is entitled Geometry with the lessons: The perimeter and The area.

5.1 Aggregated-level academic performance and learning effectiveness indicators

In this subsection, effectiveness is assessed through the score, as performance level indicator, time spent, as effort level indicator, and both, as learning efficiency indicator, in pre-and post-tests, which are summed for both units.

5.1.1 Pre- and post-test scores

Figure 11 shows that most students get higher post-test scores than the pre-test scores for both groups. Consequently, the mean of post-test scores is higher than the mean of the pre-test scores by 8.92% and 6.61% for the VARK and gamification groups, respectively.

Fig. 11
figure 11

Pre and post test scores of (a) VARK and (b) gamification groups

5.1.2 Normalized learning gain (NLG)

NLG [53] is given by:

$$ {\textit{NLG}}=\left\{ \begin{array}{ll} \frac{{Post}-{Pre}}{100-{Pre}} \times 100, \quad & \qquad {\textit{Post}} \geq {Pre}, \\ \frac{{Post}-{Pre}}{{Pre}} \times 100, \quad & \qquad {\textit{Post}} < {Pre} \end{array} \right., $$
(4)

where Pre and Post refer to the students’ test scores before and after APPEAL’s experiment, respectively; and 100 is the maximum score. NLG is averaged for the two units and plotted (in percent) for each student individually in a discrete plot as shown in Fig. 12. NLG= 100 means full mark in the post-test. NLG= 0 when both scores are equal, i.e., no enhancement in the score. It is negative when the score deteriorates. Figure 12 shows that the majority of the students achieve a high NLG; 61.54% of VARK group students achieved NLG higher than 50% and 38.46% of gamification group students achieved NLG higher than 35%. It can be inferred from Fig. 12 that, using NLG, the improvement in the scores of VARK group exceeds gamification group.

Fig. 12
figure 12

NLG of (a) VARK and (b) gamification groups

5.1.3 Pre- and post-test completion time

The test scores are not always enough to assess learning effectiveness. For example, a student may get the same score but finalizes the test in shorter time, which can be considered as an increase in learning efficiency. In other cases, students may not be careful enough in solving the post-test and finish it quickly with mistakes due to boredom or having no obligation on getting high grades. They might be only caring about finishing the experiment more than getting high grades as it is a voluntary experiment not related to the official grading system due to the exceptional circumstances of COVID-19. To overcome these obstacles, time should be considered as well.

First, we consider the completion time of both pre and post tests for both groups in Fig. 13, which indicates that 100% of the students spent shorter time in the post-test than the pre-test for both VARK and gamification groups. Consequently, the mean of post-test times is lower than the mean of the pre-test times by 54.69% and 49.68% in VARK and gamification groups, respectively. In addition, the mean of the post-test time is roughly equal for both groups, which can be considered an indication of consistency in the test completion time after APPEAL experiment.

Fig. 13
figure 13

Pre- and post-tests completion time of (a) VARK and (b) gamification groups

5.1.4 Learning efficiency (LE)

As a collective metric, a definition of LE, which includes the time spent in solving the pre and post-tests as well as the score, is also investigated. In this work and inspired from [24, 53], we define LE as Score/Time, which can be more indicative in the previously mentioned cases. Figure 14 shows the improvement of LE values by 168.59% and 149.72% for VARK and gamification groups, respectively.

Fig. 14
figure 14

Pre and post LE of (a) VARK and (b) gamification groups

These findings provide an answer to RQ1 for both groups, where APPEAL helps improving aggregated-level academic performance and learning effectiveness indicators (pre- and post-test scores, completion time and learning efficiency). The enhancement in scores, time, and accordingly LE for both groups can be attributed to a combination of several insights:

  • APPEAL experiment improves student performance in both score and time.

  • Scores in pre-test were already quite high, which indicates a high percentage of the students with advanced levels; yet, they spent longer time to solve the test. Hence, there is a higher margin and chance of enhancement in time than scores resulting in better LE.

  • Small sample size.

  • Similar ranges of time savings were obtained in [27] when comparing two systems with similar content, but different navigation strategies.

LE can be similarly explained as it is a dependent parameter. Although the mean of the LE in the post-test for gamification group is higher than VARK, the relative improvement in the mean of LE of VARK group exceeds gamification group due to having similar relation between pre-test LE means (gamification pre-test LE mean higher than VARK). Moreover, the mean NLG equals 50.64% and 35.74% for VARK and gamification groups, respectively. So, in conclusion, although APPEAL experiment helps both groups improving their academic level using different learning effectiveness measures, VARK group exhibits more improvement compared to gamification group.

5.2 Statistical analysis and data dispersion for academic performance and learning effectiveness indicators

The relatively small sample size represents a challenge against performing statistical analysis and quantifying data dispersion or variability. To further analyze the test scores statistically, box-and-whisker plot is used [46], which is more suitable than other explanatory data analysis tools for asymmetrical or highly skewed distributions with possible outliers. The spacings between the different parts of the box indicate the degree of dispersion (spread) and skewness in the data.

By comparing pre and post-tests of the VARK group, Fig. 15 indicates that APPEAL experiment helps increase the median, minimum and maximum bounds and quartiles of the student scores. In addition, APPEAL experiment reduces the students’ scores dispersion as they become more clustered around the median especially for the lower half of the scores below the median. Similarly, the median, minimum and maximum bounds and quartiles as well as the dispersion of students’ completion times decreased in the post-test compared to pre-test. Table 2 further indicates the positive impact of APPEAL experiment on score and completion time of VARK group, where the medians of scores and completion time improve by 6.06% and 53.89%, respectively. The variation between minimum and maximum (range) reduced with improvement of 58.15% and 31.18% for score and completion time, respectively. Scores and completion time in post-test are more clustered around the median, i.e., Interquartile range (IQR) reduced, than pre-test with improvement of 38.26% and 18.01%, respectively.

Fig. 15
figure 15

Statistics of pre-and post-test (a) score and (b) time for VARK group

Table 2 Boxplot summary and relative improvement of pre- and post-test score and time statistics for VARK group

Similar level of improvement in the statistical properties is noticed while studying the impact of the APPEAL experiment on the gamification group. Figure 16 and Table 3 can be described similarly achieving an improvement of 3.18% and 62.67% in the medians of score and completion time, respectively. The range reduced with improvement of 64.43% and 35.77% for score and completion time, respectively. The IQR reduced with improvement of 67.47% and 38.26%, respectively.

Fig. 16
figure 16

Statistics of pre-and post-test (a) score and (b) time for gamification group

Table 3 Boxplot summary and relative improvement of pre- and post-test score and time statistics for gamification group

These findings provide an answer to RQ2 for both groups, where APPEAL helps improving the data dispersion for the academic performance and learning effectiveness in terms of score and time.

5.3 Disaggregated-level academic performance and learning effectiveness indicators

It was found that learners with low prior knowledge can benefit much more than those with high prior knowledge [54, 64]. Around 60% and 76.93% of the students in VARK and gamification groups, respectively, got high scores in the pre-test of both units and, hence, started at the advanced level or hard exercises already. These advanced level students preserve their level. Table 4 gives a list of the students who got pre-test scores less than 75%, which represent 30.77% from the total number of students in VARK group and 23.08% from gamification group. It can be inferred from Table 4 that the post-test level of 100% of these students improves than their pre-test level. In addition, 100% of the students reached the advanced level and became capable of solving hard exercises correctly. Table 5 presents examples extracted from the exercises log data showing how these students’ performance gradually improves in different taxonomies and lessons. These are the real scenarios corresponding to the hypothetical scenarios of Section 4.3. These findings provide an answer to RQ3, where APPEAL helps improving disaggregated-level academic performance and learning effectiveness indicators for each student on lesson and exercises level.

Table 4 Students level improvement summary
Table 5 Performance improvement examples from exercises log data

5.4 Student engagement and satisfaction indicators

APPEAL survey tests the satisfaction level among the students and assesses the different hypotheses of the project from their perspective. It consists of 11 questions, which have graduated choices on LIKERT scale. Table 6 shows the students’ responses to the different hypotheses. The green color corresponds to strongly agree and the red color corresponds to strongly disagree. It can be inferred that there is 95% and 85.9375% overall satisfaction about APPEAL in VARK and gamification groups, respectively. Table 7 provides a detailed comparison, where 94% of the VARK group and 75% of the gamification group were satisfied by the presentation style and engagement. 95% of the VARK group and 87.5% of the gamification group were satisfied by the exercises scaffolding, navigation and feedback. Combining these two dimensions results in 94% VARK satisfaction compared to 75% gamification satisfaction. The less satisfied percentage of students mostly selected “neutral” and only around 6.25% of the gamification group were unsatisfied by presentation style and engagement and 2.5% were unsatisfied by exercises scaffolding, navigation and feedback. For VARK group, only 4% were unsatisfied by exercises scaffolding, navigation and feedback.

Table 6 Satisfaction survey results for (a) VARK and (b) Gamification groups
Table 7 Comparison of the satisfaction level for both groups

These findings provide an answer to RQ4, where APPEAL achieves good student engagement and satisfaction indicators. At first, it was expected that gamification for third grade students may yield better results by increasing engagement and satisfaction. However, the learning effectiveness measures of Section 5.1 and satisfaction survey results come in favor of VARK. This may be attributed to the initial learning styles of gamification group, which reports more than 30% for Read/Write learning style, while gamified content requires more interaction and kinesthetic abilities.

The conducted experiment comes in favor of offering different/mixed presentation modalities according to AI-based update of learning style and dynamic preferences than offering gamified content. From the relative improvement in test score, completion time, NLG and LE and satisfaction level, it can be inferred that VARK approach is slightly more suitable to achieve higher learning effectiveness compared to gamification. For both groups, the adaptive personalized exercises navigation played an important role in achieving this relative improvement and high satisfaction.

6 Conclusions

This paper proposed an AI-based Adaptive Personalized Platform for an Effective and Advanced Learning (APPEAL) for school students, which achieves more effective learning with reduced time spent, improved grades and satisfaction level. APPEAL is built as a dedicated website based on Moodle including a front-end interface, back-end algorithms and API. APPEAL provides user accounts with different roles (student and teacher). A dynamic learner model was built including cognitive (learning style); behavior (effort, support and performance) and affective (engagement) categories of the student personal traits. APPEAL presents a new combination of the following:

  • VARK adaptive personalized content presentation on one hand or gamification on the other hand.

  • Adaptive personalized content presentation employs a DQN RL AI implementation.

  • Adaptive personalized exercises difficulty scaffolding and navigation (through skipping/hiding less difficult exercises) and adaptive feedback (through hints and messages), as well as reattempting exercises till mastery employing an online rule-based decision making based on student interactions.

The performance of APPEAL was experimentally tested in regard to the four research questions and proved to achieve an improvement in learning effectiveness in a real life pilot experiment:

  • For RQ1, data analysis demonstrated that APPEAL experiment has a positive impact on the performance of all participants in the post-test compared to the pre-test. An improvement in the mean of test scores (9% VARK and 7% gamification), completion time (55% VARK and 50% gamification) and LE (169% VARK and 150% gamification) were reported. In addition, mean NLG of 51% and 36% were reported for VARK and gamification groups, respectively.

  • As for RQ2, around 31% from the total number of students in the VARK group and 23% from the gamification group started the experiment with easy and medium levels in the pre-test. APPEAL helped 100% of these students to enhance their performance and reach hard (advanced) in the post-test.

  • Statistically, APPEAL experiment improves the median and reduces the dispersion, which provides an answer to RQ3.

  • Regarding RQ4, 95% and 86% satisfaction level were reported for VARK and gamification groups, respectively.

  • Irrespective of the initial learning style from the questionnaire, APPEAL experiment, with the adaptive personalized RL algorithm that allows personalized choices of presentation styles, helped the students improve their academic performance to higher levels.

  • Both VARK and gamification groups witnessed academic and satisfaction level improvement after APPEAL experiment.

  • More improvement was reported in adaptive personalized VARK presentation group, where interactive activities and games in the kinesthetic presentation can also provide engagement, while keeping other presentation styles available, when needed.

The obtained results suggest an improvement in the post-test score (academic performance), time (effort) and hence efficiency compared to the pre-test, promising NLG values, statistics and satisfaction level for both groups, especially VARK. However, these results should be cautiously taken as the sample size was not as big as planned due to the impact of COVID-19 pandemic and social distancing procedures on schools. The number of students who voluntarily started and completed the experiment was 26 students, coordinating with their school. Yet, the motivation factor of relating their performance to a typical grading system was absent. Given the duration of this pilot experiment (2 months), the results seem reasonable, carry a good potential and encourage a future extension.

The different presentation styles offered by APPEAL can support students with physical and cognitive disabilities such as aural or visual processing deficits. In addition, the following items can be considered for future work:

  • The data collected from real students can be used to enhance the training of the model and implement another supervised model as a benchmark for comparison.

  • An experiment involving a control group with conventional classroom teaching against experimental groups, with larger number of participants, can be set up given previous history about the students who will take part in the experiment. We anticipate that the results reflect similar improvement, yet, with more variations among different students’ levels.

  • Dual or compound action RL for both presentation and exercises navigation can be addressed.

  • Deeper aspects of personalized gamification [31] can be investigated and combined in the AI implementation aspiring for better results for gamification group.

  • Multi-task DQN RL can be applied to predict both VARK and difficulty optimal actions.

  • Multi-agent RL can be applied in Computer-Supported Collaborative Learning (CSCL) for group learning purposes.

  • APPEAL can be extended to other age groups.