Introduction

Success in school is dependent, to a great extent, upon the development of skills in reading and writing gained during the early childhood years (NAEYC 1998). The development of literacy skills should therefore be nurtured from an early age. Accordingly, an increasing number of parents and other educators now hold the opinion that a child should be able to learn to read and write in kindergarten. In response, many kindergartens have adopted formalized methods of instruction for teaching literacy. However, this approach has left little room for young children’s natural and playful way of learning (Bassok et al. 2016). This has led to a call for a design approach that supports literacy development and greater child engagement in the learning processes. One response to this call is a focus on storytelling. Storytelling is a natural way of communicating with young children.

Digital technology has also become a part of most children’s everyday experiences. By the time children reach kindergarten, they are likely to have had countless encounters with various digital forms of communication. Therefore, for today’s young children, to be literate should also include developing a range of digital skills and knowledge. Digital literacy development has received far less attention than traditional forms of literacy. The increasing integration of digital forms of reading and writing into everyday life suggests that the role of technology for (digital) literacy development of kindergarten children is also a timely and important topic.

The main aim of the present study is to describe a series of activities to support literacy and digital literacy skills in which storytelling is infused within an instructional design framework. In addition, a set of literacy and digital literacy measures were developed to assess the effects of these activities. To our knowledge, the effort to create a series of activities that blend a more formal type of instruction with storytelling and to assess the effects on digital literacy skills in early childhood education presents a relatively novel approach.

Storytelling and Digital Storytelling

Storytelling is a process in which someone (the storyteller) uses vocalization, narrative structure, and mental imagery to communicate with an audience (the listener) (Peck 1989). The listener also uses mental imagery and communicates back to the storyteller primarily through body language and facial expressions (Roney 1996). This interaction is repeated in multiple communication cycles during any storytelling event. Storytelling thus becomes an act of mutual meaning-making and learning by all participants (Katuscáková and Katuscák 2013). Storytelling supports literacy development because it allows children to also hear models about how language can be used. Storytelling can also promote writing skills by encouraging children to create their own stories, modify stories that they have heard, and even write plays based on familiar tales (Cassell 2004; Nicolopoulou et al. 2015).

For a long time, storytelling was employed spontaneously rather than as a deliberate and planned instructional approach in early childhood education (Coskie et al. 2010; Phillips 2013). The use of storytelling as an engaging and meaningful teaching approach in early literacy education began with the work of Kieran Egan (1985, 1986). Egan proposed that teaching is best shaped in story formats because storytelling stimulates children’s imagination which is a very powerful learning tool. Agosto (2016) noted that storytelling can nurture cognitive engagement, critical thinking, and story sequencing. Her research also showed that follow-up activities such as discussion, retelling, and topic-related activities (written, drama, oral) can further enhance literacy development.

Storytelling has always been at the core of human activity (Lambert 2010). Individuals and societies have continuously explored new ways to make stories compelling, moving, empowering, and everlasting. Recently, this has occurred by integrating information and communication technologies (ICT), yielding a form of digital storytelling (Brígido-Corachán and Gregori-Signes 2014). Generally, digital storytelling revolves around presenting short, personal narratives (Meadows 2003) that combine images with text, narration, voice, and music (Robin 2008). In other words, digital storytelling offers children story content in digital, technology-based formats.

Porter (2005) explained digital storytelling as integrating the ancient art of oral storytelling with an array of technical tools to present personal tales with images, graphics, music, and sound, including the storyteller’s voice. This definition resembles that of Robin (2008) and Psomos and Kordaki (2015) who characterized digital storytelling as traditional storytelling with multimedia. Similarly, Kearney et al. (2011) describes digital storytelling as a short narrative captured in video format. Several studies with older children have shown that digital storytelling can generate children’s interest and learning motivation and facilitate their understanding of complex subject matter (Robin 2008; Sadik 2008).

Empirical studies on digital storytelling in early childhood education have concentrated on the teacher (e.g., Kildan and Incikabi 2015; Yuksel-Arslan et al. 2016). Effects of digital storytelling on children have, to our knowledge, been investigated for math or ICT, but not literacy (e.g., Preradovic et al. 2016). The relationship of digital storytelling to children’s overall literacy development in early childhood education is therefore yet to be explored.

Measuring Literacy and Digital Literacy Development

Literacy is the use of social practices of creating and interpreting meaning through text (Kern 2000). A person who focuses on text comprehension is reading to learn, and while doing so becomes literate. To be able to read to learn, children should learn to read (Robinson et al. 2013). There are considerable debates about when children should develop their reading skills and how the acquisition of literacy should manifest itself in the early years (Fletcher-Flinn 2015).

In early literacy development, a distinction is often made between code- and meaning-related skills (Lonigan et al. 2011). Code-related skills include print knowledge, alphabet knowledge, and phonological awareness, among other things (Owodally 2015). Meaning-related skills include vocabulary, grammatical ability, and oral narrative ability (Westerveld et al. 2015). Knowledge about causes, correlates, and predictors of children’s reading successes and failures in primary and secondary education has expanded greatly in recent decades. This knowledge has led to the development of standardized and nonstandardized methods of measuring literacy development in early childhood (Lonigan et al. 2011).

A common form of nonstandardized methods for measuring literacy development revolves around teacher observations and related assessments such as checklists, rating scales, and portfolios of children’s products. Since the procedures are not standardized, the demonstration of children’s skills may not be uniformly measured across all children. A common critique of nonstandardized methods is that teacher observations are often informal rather than well-structured, resulting in measurement of skills that only reflect the teacher’s judgment of the child (Brown and Rolfe 2005).

A popular standardized method for literacy development assessment is the dynamic indicators of basic early literacy skills test (DIBELS; Kamii and Manning 2005). The DIBELS mainly focuses on code-related skills. It consists of the following subtests: initial sounds fluency, letter naming fluency, phonemic segmentation fluency, nonsense word fluency, and word use fluency. The suitability of the DIBELS, and other standardized instruments, for measurement of early childhood literacy development, has generated considerable debate (Myers et al. 1996; National Research Council 2000). Concerns have been voiced about time required for such assessment, distractions that interfere with accurate measurement of capabilities, and young children’s limited test-taking abilities to assess competence. In view of these concerns, this study has adapted content from standardized tests to align with the directions of the local curriculum. The measurement of skills in this study focused on just three key skills of literacy development: alphabet recognition, phonological awareness, and print awareness. Meaningful items were created for authenticity of the child assessments.

The increasing use of technology in children’s everyday life has stimulated the emergence of digital literacy which has also been referred to as computer, technology, information, media, and communication literacies (Martin and Grudziecki, 2006). All these terms share the view that they involve technology other than text. Throughout this paper, we use the term digital literacy to refer to the use of social practices of creating and interpreting meaning through texts with technology (e.g., Kern 2000). According to Ng (2015), a digitally literate person is a competent user of three dimensions of digital technology: technical, cognitive, and socio-emotional. The dimension of technical skills is the most developed area of digital skills measurement. It includes knowledge of devices and operating skills with a focus on technical usage of computer (Ba et al. 2002), online abilities (Sonck et al. 2011), exploring tablets (Marsh 2016), and operational use of digital devices (Eshet-Alkalai and Chajut 2009; Eshet-Alkali and Amichai-Hamburger 2004; Eshet 2012). However, there is little research related to the assessment of cognitive or socio-emotional dimensions of digital literacy.

The present study considers cognitive and socio-emotional skills dimensions because they are important prerequisites of digital literacy. This choice is also in line with a definition of digital literacy in early childhood education that has emphasized the use of digital and non-digital practices in using different digital technologies (Burnett et al. 2014; Sefton-Green et al. 2016). The cognitive skills dimension of digital literacy includes critical thinking and multimodality, which infers that communication and representation is more than just about language. The socio-emotional skills dimension includes communicative and social skills. We did not assess technical skills development in this study because available tests are too complex for young children, and we wanted to stay away from highly specific facets of technical skills and investigate more general cognitive and socio-emotional dimensions.

Measurement for both literacy and digital literacy development in the present study has taken into account two critical considerations: (1) a focus on content and determining what should be measured, in order to be able to assess the effectiveness of designed activities; (2) establishing a procedure which was not time-limited in order not to put children in stressful situations.

The Current Study

The main research question addressed in this research is: Do oral or digital storytelling activities increase literacy and digital literacy development? In order to answer this question, an experimental design was adopted that included three different conditions: a control condition with usual literacy practices implemented in one classroom and two experimental conditions implemented in two other classrooms. One condition focused on literacy through oral storytelling, and the other condition focused on digital literacy using technology-based storytelling. No differences in children’s outcomes for literacy were expected between the two experimental conditions.

Method

Participants

The study was conducted in a public preschool in Indonesia. The study involved three classrooms that accommodated 45 children (25 girls and 20 boys), with an average age of 5.39 years (SD = 0.41). All classrooms used the same curriculum.

Research Design

The research design was quasi-experimental with three conditions. Classrooms were randomly assigned to conditions: control (C); oral storytelling (S); and digital storytelling (DS). In the control condition (C), the children received their regular classroom literacy enhancement activities. This condition yielded baseline data and information on maturation of literacy and digital literacy development across the time of the study. In one experimental condition, the children received oral storytelling activities (S), in which a story was read aloud and presented with some visual clues. In the other experimental condition, the children received digital storytelling activities (DS) and the children watched and listened to a recorded story.

Instructional Materials

There were three storytelling sessions for each of the experimental conditions. The theme for the three storytelling sessions was “Me.” This is a common theme for classroom activities at the beginning of the school year. The sessions focused on “My Name,” “My Birthday,” and “My Hobby.” The content of sessions had an identical story line across the experimental conditions. The didactic approach in each activity session followed Gagné’s events of instructional design (Smith and Ragan, 2005). For example, the theme of the first activity was “My name” (See Table 1). The main objective was to enhance the children’s understanding of their own identity as presented in oral and printed form. More specifically, the activities supported the development of the recognition of letters and the sounds associated with letters in children’s full name and nickname. The story line revolves around a boy who likes his name, recognizes his name written on his belongings, and also recognizes the sounds of his name.

Table 1 Gagne’s events of instruction for the activities focused on the theme, “My name”

Assessment Instruments

Literacy Measures

The literacy measures assessed three core skills of early reading and writing, namely alphabet knowledge, phonological awareness, and print awareness. Alphabet knowledge and phonological awareness play different roles in literacy acquisition and development. These two sets of knowledge are both necessary for the acquisition of literacy (Muter 1994). The same literacy measures were used before and after the children had participated in the three intervention activities.

Alphabet knowledge is one important aspect in the acquisition of literacy during the early childhood years (Foulin 2005; Wood and McLemore 2001). It is defined as the ability to name, distinguish shapes, write, and identify the sounds of the alphabet (Piasta and Wagner 2010).

Phonological awareness is another important skill linked to the acquisition and development of reading skills in school (McLachlan and Arrow 2010). It is an awareness of sounds in spoken words that is revealed by abilities such as rhyming, matching initial consonants, and counting the number of phonemes in spoken words. Phonological awareness has been measured, and consequently defined, by many different tasks. One of them is isolating single sounds from words (Murray et al. 1996).

Print awareness refers to children’s ability to recognize the function and form of print and the relationship between oral and written language (Pullen and Justice 2003). Important reading prerequisites include children’s ability to recognize environmentally embedded and contextualized print, to understand the form and function of print, and to perceive the relationship between speech and print (Kassow 2006). In their study, Justice and Ezell (2001) highlight the importance of measuring this ability by using storybooks in order to provide a meaningful context for the words.

Literacy Assessment

This involved assessment of five skills: (1) identify the initial sound of words (phonological awareness), (2) know your own name (print awareness), (3) recognize letters (alphabet knowledge), (4) recognize everyday words (print awareness), and (5) recognize names in written form (print awareness). The child’s skill level was measured by asking increasingly more complex questions for each feature. Assessment of any of the literacy skills was stopped if a child failed to correctly answer a question. The children’s responses were scored with a rubric using a 4-point scale for each task (See “Appendix A”). Before beginning each part of the literacy assessment, the child was given two practice items.

For example, the process for the literacy assessment item of “identify the initial sound of words” involved stimulus material consisting of 12 words in everyday use, and each word consisted of two to three phonemes. After hearing the word, the child was asked to say the first sound in the word. Because many children would identify the name of the letter rather than the sound, they were given feedback on their responses to the practice items to make sure that they understood the task. If a child failed to identify the initial sound of the first word given, he/she would receive no points for that part of the assessment. If a child answered the first item correctly, a new word was then presented. Reliability analyses using Cronbach’s alpha showed that there were satisfactory results for the overall literacy pretest (α = 0.63) and posttest (α = 0.79).

Digital Literacy Measures

Measurement involved an assessment of cognitive and socio-emotional skills that are considered prerequisites of digital literacy. Since this procedure was supplemental to, and independent of, the general classroom literacy assessment, data on these measures were gathered from five randomly chosen participants from each classroom (i.e., five children per condition). The average age of these 15 children (eight girls and seven boys) was 5.47 years (SD = 0.33) which was similar to the average age for the whole sample.

Cognitive skills items assessed early expressions of multimodality and critical thinking, abstract and reflective thinking (Kazakoff 2015; Wenner et al. 2008). The measure linked the items to storytelling and operationalized these items as recall of a past event, planning for a new event and picture reading. For instance, recall of a past experience is important for story comprehension. It hinges on reflection about the context of an event (who, where, and when) and on the ability to explain an unfolding story (how and why). The ability to “read” a picture is an important skill in digital literacy. This facet in the digital literacy measure was the counterpart of the purely text-oriented items in the literacy assessment.

Socio-emotional skills were assessed by communication about the self with others through digital platforms (Ng 2015). An important aspect of this skill in early childhood concerns is having a sense of self-identity (Marsh et al. 2005). The development of self-identity includes knowing one’s name, age, and gender. In addition, it concerns how the child’s understands his/her place in the world (Raburu 2015).

The digital literacy measure consisted of a rubric with the following items: (1) recall a past event (cognitive skills); (2) plan an event (cognitive skills); (3) read a picture (multimodality—cognitive skills); (4) understand one’s own identity (socio-emotional skills); and (5) engage in a conversation (socio-emotional skills).

Digital Literacy Assessment

Administration of the digital literacy measures followed the same procedure as for the literacy test (see “Appendix B” for all items and the scoring rubric). For the measure “recall a past event,” the items were focused on children’s ability to identify details from a single, personal experience. Each child was asked to recall a recent holiday experience. If a child could not recall a single experience, the score for this test item would be zero points. If a child recalled an event, he/she was asked to provide details about that experience. The allotted points would then depend on the number of details that were given (i.e., what, where, when, with who, and how). If a child started to mention another event, the experimenter would redirect the discussion to the original event. If the child could not do this or could not give details, the questions for this item ended. Reliability analyses using Cronbach’s alpha showed that there were satisfactory results for the digital literacy pretest (α = 0.65) and posttest (α = 0.87).

Procedure

The intervention study consisted of three phases: pretest (Week 1), implementation (Weeks 2–4), and posttest (Week 5). In the pretest, a group of three to five children gathered in the reading room of the school where the experimenter and a research assistant were present. The experimenter would then engage with each child, in turn, for the assessments. Administration of the literacy assessment took 10–15 min for each child, while the digital literacy assessment took 25–35 min for each child. During the assessment process for each child, the other children in the room engaged in playful activities led by the research assistant.

During the implementation phase, in the control condition, the children participated in the regular literacy-focused classroom activities led by the teacher and with the experimenter present. In the experimental conditions (storytelling and digital storytelling), the activities were given by the experimenter while the regular classroom teacher was present to assist with the children. The overall structure of these activities was the same in both conditions (see also Table 1). First, there was storytelling. Next, there was a whole-class discussion about the story, and finally, there were follow-up activities. Each storytelling session began with the experimenter explaining the rules for the session and preparing the children for the story by informing them about the title, characters, and main idea.

In the oral storytelling condition, the experimenter then read the story along with presenting some visual clues to the children. While in the digital storytelling condition, the story was prerecorded and included texts, pictures, voices, and sounds and then was digitally presented using a projection device. The story in the first storytelling session was about a boy who likes his name, followed by a whole-class discussion about the story. This was followed by two story-related activities. One of these was an individual project with a focus on children’s creativity; for example, children designed a book label using the letters of their own name. The second follow-up activity was a group project, such as dramatic play involving an exchange of the children’s name cards.

The posttest was administered in the week following the three activity sessions with a similar procedure to the pretest sessions.

Data Analysis

Eight of the 45 children participating in the intervention study did not complete both the literacy pretest and posttest because of absence (five children) or because they were special needs students (three children). Only complete data for the 37 children who completed all assessment tasks were used in the final analyses.

Because the distribution of scores on some measures in the pretest and posttest violated measurement assumptions for normality and homogeneity of variance, we report the findings using nonparametric tests (i.e., Kruskal–Wallis H test). Significant findings were followed by post hoc tests (i.e., Mann–Whitney U test); two-tailed tests were used with alpha set at 0.017 (0.05/3), using Bonferroni corrections for multiple tests. For effect size, we report the r-statistic (Field 2005). This statistic is qualified as small, medium, and large effects for the values r = 0.10, r = 0.30, and r = 0.50, respectively. For change in assessment from pre to post, key assumptions for gain scores (i.e., posttest score minus the pretest score) were met to assess change using analysis of variance (ANOVA), followed by Tukey HSD post hoc analysis after significance. For these effect sizes, we report the d-statistic (Cohen 1988). The findings are classified as follows: small for d = 0.20, medium for d = 0.50, and large for d = 0.80, respectively.

Results

Development of Literacy Skills

Table 2 presents the mean item scores and standard deviations for the literacy and digital literacy assessments for each group (control condition; oral storytelling condition; and digital storytelling conditions). These mean scores showed that the overall scores of the participants in all three conditions at pretest and posttest were below the mid-scale value of 2. In addition, there were moderate levels of variance, as indicated by the standard deviations.

Table 2 Descriptive statistics for literacy assessment for each condition before and after the intervention activities

The scores on the literacy pretest did not differ between conditions, H (36) = 0.648, p = 0.723. In contrast, there was a statistically significant difference on the literacy posttest between conditions, H (36) = 8.26, p = 0.016. Post hoc tests showed that there was a statistically significant and large effect for the comparison between the control condition and digital storytelling condition, U (25) = 128.50, z = 2.82, p = 0.005, r = 0.56. There was no difference between the control and storytelling condition or between the storytelling and digital storytelling condition, respectively, U (25) = 81.50, z = 0.193, p = 0.847; U (25) = 90.50, z = 0.485, p = 0.485.

The boxplots for the literacy posttest, shown in Fig. 1, provide a visual presentation of differences in outcomes by condition. The shaded areas in each box represent the mid 50% of the scores. The horizontal line in the box is the median score in each condition. The top (or bottom) 25% of scores are shown in the distance between the highest (or lowest) horizontal line and the highest (or lowest) edge of the shaded box. The shaded boxes of the control and digital storytelling condition do not overlap (indicating significant difference between these two groups as identified by the statistical tests). In contrast, there is some overlap between the oral storytelling and digital storytelling conditions. The median scores of the two experimental conditions are also similar.

Fig. 1
figure 1

Boxplot for the literacy posttest scores for each condition

Data analyses for the learning gain (Lgain) scores showed a mean overall improvement for literacy skills of 0.21 (SD = 0.21). For the control, storytelling, and digital storytelling condition, the mean Lgain score was, respectively, for each condition, M = 0.03 (SD = 0.08), M = 0.20 (SD = 0.19), and M = 0.38 (SD = 0.18). The ANOVA for Lgain scores showed that there was a statistically significant difference between conditions, F (2, 33) = 14.74, p < 0.001. Post hoc comparisons indicated that the control had a lower Lgain score than the storytelling condition, p = 0.014, d = 1.17, and also a lower Lgain score than the digital storytelling condition, p < 0.001, d = 2.53. For both comparisons, a large effect size was obtained. In addition, the analysis showed that there was a significantly higher Lgain score in the digital storytelling compared to the storytelling condition, p = 0.008, d = 0.97, and with a large effect size.

Development of Digital Literacy Skills

The pretest results showed that all participants (n = 15) in the three conditions (control, oral storytelling, digital storytelling) had comparable digital skills at the start of the activities, M = 1.88 (SD = 0.23), M = 1.92 (SD = 0.18), and M = 1.88 (SD = 0.23). After the activities, the posttest results were M = 2.04 (SD = 0.22), M = 2.80 (SD = 0.28), and M = 3.00 (SD = 0.42). The results also showed that the median score and the distribution of the digital literacy posttest scores in the control condition did not overlap with those of the experimental conditions. In contrast, there was an overlap of the distribution of the scores between the two experimental conditions and also the median scores were close to each other.

The Kruskal–Wallis test showed that there was no statistically significant difference between conditions on the digital literacy pretest, H (15) = 0.153, p = 0.926. In contrast, a statistically significant difference was found on the posttest, H (15) = 9.40, p = 0.09. Post hoc tests showed that there was a statistically significant and large difference for the comparison between the control and storytelling condition, U (10) = 24.50, z = 2.58, p = 0.008, r = 0.82, and for the comparison between the control and digital storytelling condition, U (10) = 25, z = 2.65, p = 0.008, r = 0.84. The two experimental conditions did not differ significantly, U (10) = 15.50, z = 0.64, p = 0.523.

Data analyses for the digital learning gain (DLgain) scores showed that all conditions showed a mean overall improvement for digital literacy skills of 0.72 (SD = 0.45). For the control, storytelling and digital storytelling condition, the mean DLgain was, respectively, M = 0.16 (SD = 0.09), M = 0.88 (SD = 0.18), and M = 1.12 (SD = 0.23). The ANOVA for DLgain showed that there was a statistically significant difference between conditions, F (2, 14) = 40.70, p < 0.001. Post hoc comparisons indicated that the control had a lower DLgain than the storytelling condition, p < 0.001, d = 5.06, and also a lower DLgain than the digital storytelling condition, p < 0.001, d = 5.50. For both comparisons, a large effect size was obtained. In addition, the analysis showed that there was no significant difference for DLgain between the digital storytelling compared to the storytelling condition, p = 0.051.

Discussion

Increased attention to (digital) literacy goals in early childhood education has been accompanied by more formalized educational practices. These endeavors have been criticized as being out of tune with how young children (like to) learn (Miller and Almon 2009). The present study investigated the design and effectiveness of a storytelling approach to achieving digital literacy goals. The storytelling activities were systematically structured by adopting the events of instructional design proposed by Gagné (Smith and Ragan 2005).

The core content of these activities involved storytelling to enable a natural, playful way of learning. This also is in accord with the view that storytelling is a cornerstone of digital literacy development. Storytelling provided the objectives, models, and motivation for literacy skills application. For example, storytelling in the first activity revolved around a child’s name and included how to say, spell, and write that name. Follow-up activities provided opportunities for practice of the presented skills. These activities stimulated the children to apply what they had learned to their own situations. This format was the same across the two experimental conditions, although in the digital storytelling condition the story was prerecorded, using multimedia including texts, pictures, voices, and sounds. In the design of the stories, special attention was given to considerations from multimedia learning theory (Mayer and Moreno 2003). That is, the design aimed for a strong match between the verbal and nonverbal information in the stories (Takacs et al. 2015). In the digital storytelling activities, the digital content aimed to support the children’s activities and engagement with the real world was highlighted in accord with the statement published by NAEYC and Fred Rogers Center (2012).

The measures used assessed a broad spectrum of literacy skills in early childhood. The use of rubrics to implement authentic assessment tasks (Bagnato et al. 2014) also avoided subjectivity in the measurement processes by the use of the rubric.

The administration of a pretest afforded an assessment of learning gains and enabled us to test the comparability of the children’s starting levels of (digital) literacy skills across classrooms. Pretest scores indicated that there were no differences in the mean skills levels in classrooms/conditions at the pretest assessments.

The digital storytelling activities had a significant effect on the children’s literacy skills development. Whereas the children in the control condition showed little improvement across the intervention period, the children in the experimental conditions achieved significant gains in literacy skills. Findings suggested that the digital storytelling activities also enhanced the children’s digital literacy skills. For digital literacy development, significant gains were found only for the experimental conditions. Storytelling activities seemed to be equally effective, across the experimental conditions, for developing the children’s digital literacy skills.

The findings showed that the effects of the digital literacy activities on digital literacy development were promising. However, this part of the research was exploratory and involved only a small sample of children who participated in the assessments. It is recommended that future research include bigger sample sizes to form more definitive conclusions on the effects of storytelling sessions on digital literacy development. A second methodological limitation of the research was that the storytelling activities were led by the experimenter while the regular activities in the control condition were taught by the classroom teacher. During the activities, the teacher or experimenter served as a teaching assistant and vice versa. We decided to adopt this course of action to disrupt as little as possible the normal classroom routine in the control classroom.

Conclusions

Overall, the research indicated that the integration of a storytelling approach with the didactic approach of Gagné’s nine events of instruction is a promising approach for enhancing literacy and digital literacy development in early childhood classrooms. Nevertheless, to move beyond the entertainment or novelty effect that digital storytelling might bring (Campbell 2012), future studies might consider designing a more extensive series of activities to reduce the novelty effect and to maximize all possible benefits that might result from digital storytelling activities.