Influencing Adoption of Evidence-based Practices in Childcare Using Feedback, Goal Setting, and Text Messaging

High-quality childcare — as measured by the degree of emotional support and environmental stimulation, safety, and protection and promotion of physical health — is associated with better child development outcomes (Burchinal et al., 2010; Bustamante et al., 2022). Elements of high-quality care associated with children’s school readiness and long-term development include structural elements, such as teacher/child ratios (NICHD, 2002) and teacher qualifications (Manning et al., 2019); and process elements (Ulferts et al., 2019), such as teacher/child interaction (Cash et al., 2019), focused attention to literacy and math (Jenkins et al., 2018), and children’s access to materials (Mashburn et al., 2008) that ensure environments are stimulating and responsive to individual children’s needs. High-quality childcare is especially important for young children who have experienced environmental risk factors such as growing up in poverty, experiencing migration, high levels of family stress, and exposure to community stressors like violence and racism (Bustamante et al., 2022). Yet in the United States, a large-scale study of early childhood settings found that less than 10% of childcare settings met standards for high-quality care (Griffin & Friedman, 2007) and more recently, analyses suggests that many early childhood settings meet slightly more than half of desired quality benchmarks (Slicker et al., 2023), underscoring the importance of focusing on quality in childcare settings.

Research has demonstrated that adopting evidence-based practices (EBP) in childcare can lead to positive impacts on observed childcare quality and child development (Egert et al., 2018). A key question facing the field of early childhood is how to promote reliance on EBP among all early childhood professionals, especially when access to professional development (PD) is uneven. Many early childhood professionals are highly dedicated to children and families. As a group, childcare professionals express motivation to stay in the field independent of compensation (Torquati et al., 2007), demonstrating the depth of commitment that many professionals feel towards children and families (Park et al., 2022). Therefore, the lack of reliance on EBP is perhaps most accurately viewed as the result of barriers such as lack of knowledge of what to do, inability to enact EBP based on lack of materials, time or other factors, and lack of belief that a particular EBP is important. Focused interventions specifically targeting barriers to adoption of EBP therefore may be effective. This study addresses the need for evaluation of scalable, cost-effective interventions to promote use of EBP in childcare settings, using mechanisms such as text messaging.

Data are limited on cost-effective approaches to encourage behavior change and increase the use of EBP. Most efforts to improve quality are delivered through PD programs which take several forms, including in-person training, coaching, and peer mentoring (Schachter, 2015). While PD varies in both form and impact, overall, research has found that targeting quality practices through PD leads to positive impacts on children’s development (Egert et al., 2018). Provision of technical assistance and high-intensity PD positively affect childcare quality in family childcare homes (Han et al., 2021; Tang et al., 2020). Coaching, based on ongoing interactions between a coach and a childcare professional, has been documented as one of the most effective pathways towards childcare quality improvement (Yang et al., 2022). Meta-analyses showed that in-person, relationship-based coaching for 45–60 h total had the largest effect sizes on improving childcare quality, but such coaching interventions are expensive and difficult to scale (Ackerman, 2008).

Another key mechanism for supporting quality improvement is giving targeted, explicit feedback on specific changes to make in early education programs (Egert, 2018). Many childcare facilities experience observations of program quality by outside observers through state quality rating and improvement systems, childcare PD programs, and other mechanisms. The purpose of these observations is to document strengths and areas for improvement, which in turn is hypothesized to lead to changes in childcare quality. Several studies have documented the value of video feedback for improvements in the sensitivity and responsiveness of childcare providers (Groeneveld, 2016; Werner et al., 2018). But there are at least two limitations to the present system of providing observational feedback that should be addressed: (1) observations are costly to scale, requiring trained observers and the willingness of early educators to allow them into early childhood settings; and (2) feedback from observations is not always provided to educators, and when it is it is not consistently provided in ways that are specific and immediately actionable, often focused on adherence to state standards rather than childcare providers’ personal goals. For example, in K-12 school settings, observations conducted by principals do not consistently lead to sustained improvements in classroom quality (Stecher et al., 2018), in part due to the lack of specific, actionable feedback and the amount of time required for principals and other school leaders to document practices well. However, less work on the value and potential impact of specific, actionable observational feedback has been focused specifically on early childhood settings.

Further, we know little about the impacts of light-touch, technology-enabled approaches to PD in childcare settings. Integrating scientifically proven behavior change techniques to encourage EBP (i.e., providing feedback to providers using objective data and encouraging behavior change through setting goals and positive reinforcement) has the potential to transform childcare environments (Carey et al., 2019) but questions remain on how to effectively deliver these interventions at scale. While tech-based interventions have been shown to promote healthy behaviors among childcare professionals (Arandia et al., 2020) less work to date has focused on interventions using behavior change techniques to promote EBP for child development. A recent systematic review reported only 11 studies with adequate rigor for judging the impacts of tech-based PD on observed quality (Snell et al., 2019). Text messaging and emails can encourage childcare providers to adopt EBP (Barton et al., 2019), although effects indicated that individual approaches were needed to maximize impact.

Insights from behavioral science have demonstrated that exposure to brief early childhood interventions, or “nudges” can lead to positive changes over time. Nudges have been defined as a method for “influencing … behavior through suggestions, positive reinforcement and other non-coercive means” (Karlsen et al., 2019; p.1). Specifically, nudge theory has posited that brief, focused messages on aspects of behavior change can lead to positive changes in reliance on EBP. Brunsek et al. (2023) recently reported positive changes in response to a brief PD intervention relying on several forms of focused feedback and text messaging. Using a text-based approach to support teachers in a large-scale study, Hanno (2023) reported nudges on two goals: making transitions more efficient and promoting language development. The system was effective in encouraging teachers to speak more to children (although they listened less) and had no impact on changes to transitions. Hanno concluded that concrete messages that are feasible to implement could be a critical piece of effective use of text messaging. Given the potential value of inexpensive, tech-enabled PD techniques for scaling quality early childhood programs, more work is needed to examine the role that nudges delivered through text messaging can play.

This study was designed to examine the impacts of a virtual observation of routine childcare practices followed by providing childcare professionals feedback on target EBP, goal setting, and text-message nudges on the adoption of specific EBP in typical childcare settings in Nebraska. Rather than an emphasis on broad, comprehensive ratings of quality, we were interested in testing the impacts of a tech-enabled system on the adoption of specific practices. Based on previous work, we hypothesized that an observation coupled with goal setting and a focus on a small set of EBP, reinforced through text-message nudges, would lead to a greater uptake of EBP than an observation and feedback alone. We also hypothesized that educators’ perceptions of the importance of specific EBP and their sense of personal efficacy would moderate results, with greater behavior change among educators who believed the EBP were important and who experienced high levels of personal self-efficacy.

Method

Participants

Prior to recruitment, the study design was approved by the Institutional Review Board. Childcare professionals were recruited to participate in the study through publicly available lists of childcare providers, childcare networks providing PD, social media, and through word of mouth among childcare providers between August 2022 and February 2023. Educators were recruited statewide, including rural areas, through childcare networks in Nebraska. Once an educator signaled interest, they were provided with materials outlining the purpose and process of the study for informed consent and were asked to complete an initial needs assessment survey.

A total of 81 educators initially responded to the study call, filled out the needs assessments survey, and agreed to participate in the full study. The needs assessment survey included questions on educators’ income, age, PD received in the past year, overall levels of stress, beliefs about the importance of quality practices, and feelings of self-efficacy. After completing the needs assessment survey, participants were contacted to schedule a 90-minute virtual observation of a typical morning in their early childhood setting. A total of 46 educators scheduled virtual observations. Prior to the observation, a pre-meeting was held to tour the program area and to ensure appropriate logistics, such as the device placement and connectivity, were in place for the virtual observation. After the observation was complete, participants were randomly assigned to treatment (n = 23) and control (n = 23) groups. At endline, there was no attrition in the treatment group versus seven of 23 participants in the control group who did not complete the final observation. Educators were compensated $50 for completing initial needs assessment and observation, and $50 for endline observations. Educators who completed the intervention received an additional $50 for participating in the text message program.

Of the 46 participants, 37% were working in childcare centers, 58% were working in licensed family childcare homes, and 4% were in unlicensed family childcare homes. Educators had been working in childcare from 1 to 40 years (mean = 15.00 years; SD = 10.71); ranged in age from 22 to 63 years; and had formal education ranging from high school to advanced degrees. Participants’ total household income ranged from less than $25,000 per year to more than $120,000 per year. Eighty-five percent of participants identified their race as white, 2% identified as Black, and 13% as another race/ethnicity. Eleven percent of participants reported that all children in their program identified as Black, Indigenous, and people of color (BIPOC), and 59% of participants reported that some of the children in their color identified as BIPOC. Family childcare home educators reported more PD hours (mean = 30 to 45 h) than childcare center educators (mean = 15 to 30 h) and were less likely to enroll children on subsidies.

There were no significant differences between treatment and control group participants on age, income, year of formal education, caring for children on subsidies, family childcare home vs. center, amount of PD received in the last year, overall stress levels or baseline quality observation scores. Treatment group participants were more likely to have received an observation of childcare quality within the past year (65% of treatment group participants vs. 35% of control group participants). Treatment group participants were evenly split between centers (n = 10) and family childcare homes (n = 10), with 2 unlicensed childcare providers. We therefore concluded that the randomization process was successful.

Measures

Needs Assessment Survey

Needs assessment survey questions included educators’ ages, level of formal education, PD received, participation in the state quality improvement and rating system, whether an observation had been completed in the past year, their race/ethnicity as well as the children they were caring for, their views on the importance of specific quality practices in childcare settings, perceived self-efficacy and stress levels. Perceived self-efficacy was measured using a validated scale (Schwarzer et al., 1995). All other questions were generated by the research team to align with study goals. Stress was measured by asking educators to rate the extent to which they were stressed by a series of job-related stressors, including addressing children’s behavior problems, dealing with administrative issues, communicating effectively with parents, enrolling children, and having enough resources for the program. Perceived importance of specific practices was measured by asking educators to rate each practice using a sliding scale of 0 to 100.

Brief Early Childhood Quality Inventory (BEQI) Observation

The BEQI tool is an observational measure of early childhood quality that is primarily comprised of binary yes/no items indexing specific EBP; use of materials is indexed using a 3-point scale (not available; available but children not observed using materials; and material available and at least one child observed using). BEQI was developed based on available literature on specific practices that promote child development followed by expert review of the items, rated for importance for child development outcomes. After this process, a total of 44 observational items were retained for the final tool. Scores are created by summing across items, with higher scores indicating that more desired behaviors are observed. Because the focus of this study was on the ability of text messaging to support changes in a small set of EBP, we selected eight BEQI items for this study. We refer to these eight selected practices as “targeted EBPs” in this paper. These practices included (1) creating the time and space for children to read books on their own; (2) educator reading story books to children; (3) talking with children about emotions; (4) working one-on-one with children; (5) connecting activities to children’s lives; (6) integrating math concepts into activities and conversation; (7) integrating literacy concepts into activities and conversation; and (8) integrating science concepts into activities and conversation. These items were identified because of their significance for child development and because preliminary data (not reported here) showed a variability in how frequently these practices were observed within another sample from the same population. See Table 1 for a summary of the literature supporting each of the selected eight practices.

Table 1 Rationale for 8 potential target practices

Observers were highly trained early childhood experts with experience conducting observations of childcare quality using several different tools. Seven observers participated in a four-hour virtual BEQI reliability training. At the end of the training, observers completed two BEQI Reliability Assignments which included a written quiz, made up of 20 multiple choice items related to the BEQI. Following the written quiz, observers watched a 30-minute video from a family childcare program and scored the video using the full BEQI Observation tool. Scores were then compared to the master coded scores. To become reliable on the BEQI, observers had to score above 85% on both assignments. At the end of the training, all seven observers passed both quizzes.

A 90-minute virtual program observation was performed for all educators. Observers completed the BEQI observation scorecard immediately after the observation was complete. To minimize bias, observers were blind to whether educators were in control or treatment group and were not aware which EBP practices were targeted or set as goals. Once the observation was complete, all participants were sent feedback, described in greater detail below.

Weekly Progress Reports Via Text Messaging

Educators in the treatment group were asked to respond to weekly surveys sent via text message with questions regarding their progress towards their selected goals for quality improvement and the barriers that stood in the way (more information on the goal setting can be found below, “Treatment Group Intervention”). They were also asked to report on how happy they were feeling in their jobs. As outlined below, text message surveys were sent to educators weekly during the intervention period. Control group participants did not receive text messages.

Feedback on Program Participation and Impacts

At the end of the intervention period, we asked treatment group participants how likely they would be to recommend the program to other educators and how much or little they believed the program impacted children in their program, along with questions about which programmatic activities were helpful to them.

Procedure

After completing the needs assessment and observation, educators received observation feedback. To minimize cognitive overload and in line with principles of behavioral science, educators only received feedback on a small number of practices. In the treatment group, educators received feedback on three “areas of goalsetting” (i.e. EBP they were not observed doing) and six “strengths” (i.e. EBP they were observed doing); in the control group, educators received feedback on a maximum of four EBP they were not observed doing and four EBP they were observed doing (with the goal of balanced feedback on observed strengths and weaknesses).

Treatment Group Intervention

First, educators received a strengths-based summary of their BEQI observation. The summary first validated the educator’s strengths (six EBP observed during the BEQI observation) and then focused on three target practices (based on non-observed EBP), which were framed to educators as “Areas for Goalsetting”. We randomized which practices we gave feedback on, prioritizing the eight target practices as mentioned above. See Fig. 1 as an example of the Observation Summary provided to the treatment group. BEQI observation summaries were sent to the educators in the treatment group via text message.

Fig. 1
figure 1

Treatment group observation summary sample

Upon receipt of the observation summaries, educators were asked to select one goal from their three identified “Areas for Goalsetting” for an EBP they wanted to address. The intervention was implemented in two rounds of six-weeks (12 weeks in total.) From the “Areas for Goalsetting,” educators chose a first goal of an EBP to focus on in the first six weeks (weeks 1–6). At the end of the first six weeks, educators selected a second goal (from the same list) to focus on for the second six weeks (weeks 7–12).

Educators in the treatment group received “Together in Practice” Kits via mail. These kits included Together In Practice (“TIP”) cards based on the specific goals they selected. The TIP cards outlined why their selected EBP was important, suggestions and examples of how to implement it, along with simple materials to support them in implementing the EBP. See Fig. 2 for an example of a TIP card. Educators received two TIP Kits: one per each selected goal for each of the six-week text goal setting program.

Fig. 2
figure 2

Together in practice “TIP” card example

Educators in the treatment group also received and responded to weekly text message “nudges” to support them with their goals. Educators received two text messages per week. Text messages included (a) weekly brief surveys to answer questions about how the educators were feeling, how often they implemented their goal in their childcare over the previous week, and any barriers towards implementing the goal; (b) weekly light PD in the form of a text message, including some basic information about why their goal was important, suggestions of how to do the goal and materials that could be used to support the goal, and video examples of other educators doing the goal.

At the end of each of the six-week goalsetting program, participants were sent a final survey via text message to reflect on their overall change in practices, their perceptions of the program in terms of helping improve general practices, and to provide feedback on the individual components of the goal setting program (i.e. receiving observation summaries, making goals, receiving text messages, receiving the TIP kit, etc.). Educators in the treatment group participated in two rounds of this six-week text messaging program; each round focused on a different selected goal. See Fig. 3 for a summary of the full 12-week text message goal setting program.

Fig. 3
figure 3

BEQI intervention summary

Control Group

Educators in the control group received a simple summary of eight practices following their initial observation. We focused on the same eight target EBP as the treatment group and randomized what feedback we provided. Educators received feedback on up to four practices they were not observed doing and at least four practices they were observed doing. If there were less than four of the target practices observed, we provided feedback on practices they did from other preselected items from BEQI observation. These basic summaries were sent via mail and no additional feedback was provided. See Fig. 4 as an example of the Observation Summary provided to the control group. The control group did not participate in a text messaging program or receive any additional support.

Fig. 4
figure 4

Control group observation example

End of Intervention for Both Groups

After an average of 15 weeks (minimum 79 days- maximum 216 days elapsed for all participants; mean 108 days elapsed for treatment group and mean 104 days elapsed for control group participants), participants in both the treatment and control group participated in another 90-minute virtual BEQI observation using the same tool as in the beginning of the intervention. A full summary of the intervention can be seen in Fig. 3.

Analytic Plan

Our main hypothesis was that educators in the treatment group would show a greater degree of improvement in use of target quality practices (e.g. specific not-observed EBP that participants received feedback on) than educators in the control group. As such, our analytic strategy was focused on comparing the control and treatment groups on use of target practices. We constructed two outcome variables as proxies for quality. First, we constructed a variable to summarize changes in EBP that were designated as “target practices” from Time 1 to Time 2 for the treatment group. This variable identifies whether unobserved practices that educators received specific feedback on were observed in the Time 2 observation, summarized as a percentage. Given that participants did not necessarily receive feedback on the same number of EBP, we calculated individual scores as a percentage (i.e., if an educator received feedback on three EBP and showed changes in two, they would receive 0.66 as their score for this variable). Second, we report on the total scores for the eight target practices at Time 1 and Time 2.

Results

Time 1 Needs Assessment Results

The goal of our analyses of Time 1 survey responses were to identify associations between educators’ education, PD, feelings of self-efficacy, and perceived importance of EBP, and Time 1 observations of the eight practices, as a first step in defining how these factors were related to intervention effects. Educators gave high ratings to the importance of EBP, with means of 85 or higher on all items. Science and math activities were given lower ratings overall than the other practices, with an average of 87 on math activities and 85 on science activities compared with averages of 97 on reading books to children and talking about feelings, which may reflect lack of support for early math and science among childcare professionals (e.g., Geist, 2015; Piasta et al., 2015). But contrary to our hypothesis, there were no significant associations between the degree to which educators rated each practice as important and whether they were observed as engaging in that practice during the observation. We also found no associations between engagement in the eight practices at Time 1 or overall BEQI scores, ratings of importance of target behaviors, and educators’ feelings of self-efficacy and stress.

Time 1 Initial Observations and Goal Setting

Total children present during the initial observation ranged from 2 to 22 (mean = 7.83; SD = 3.88); total number of educators ranged from 1 to 3 (mean = 1.15; SD = 0.42). Table 2 displays the frequency of observed target practices for both groups of educators. Of the eight target practices, the most frequently observed practice at Time 1 was educators working one-on-one with children (87%); least frequent was educator incorporating science activities (13%). Mean number of practices observed among the eight was 4.95, ranging from 1 to 8; the treatment group mean was 5.17 and the control group mean was 4.74 and the difference was not statistically significant. Educators who had received more PD in the past year were observed as having a higher mean score on the eight practices at Time 1 (F (3, 45) = 2.99, p < .05). We observed no other significant associations between educator characteristics measured in the needs assessment survey and overall scores on the eight practices.

Table 2 Frequency of practices observed at time 1 and time 2

Educators in the treatment group identified two goals based on observational feedback as described above. Selected goals were relatively evenly distributed across the eight. See Table 3 for the percentage of goals selected by educators.

Table 3 Feedback on target practices and rate of addressing feedback

Time 2 Observations

Total children present during the Time 2 observation ranged from 3 to 24 (mean = 8.77; SD = 3.78); total number of educators ranged from 1 to 3 (mean = 1.03; SD = 0.40). Consistent with Time 1, the most frequently observed activities were educators working one-on-one with children (82%) and the least frequent was educators incorporating science activities (15%).

Intervention Effects

Our primary hypothesis was that treatment group members would demonstrate greater improvements in target practices from Time 1 to Time 2 than control group members. Table 3 illustrates the rate of changes on the specific areas of feedback that educators received in their observation feedback summaries. In both groups, educators were most likely to improve the practice of working one-on-one with children (100% of treatment and control groups improved this EBP after receiving feedback on it). Across groups, educators were less likely to improve the practice of helping children read or look at books on their own (27% of treatment groups and 18% of control group improved this EBP after receiving feedback on it).

When we focused on changes in the two specific target practices that were selected as goals, significant differences between the treatment and control groups were noted (treatment mean = 0.62; control mean = 0.21; F (1, 37) = 19.48, p < .001; eta squared = 0.32). Treatment group members were three times more likely to demonstrate changes in the targeted EBP than control group members, with an average of 62% of EBP improved in the treatment group versus an average of 21% of EBP changed in the control group.

We found that 91% of educators in the treatment group who received feedback and participated in the goal setting text message program showed improvement in at least one of the targeted EBP, compared with 56% of educators from the control group who received feedback on areas for improvement but no further support. Said another way, members of the treatment group were more likely to improve on the targeted EBP identified in the observation through receiving feedback, texting, and TIP cards than members of the control group, who received information on the weaknesses with no additional support or nudging.

Notably, even in the absence of continued support, some educators from the control did show improvement on some of the EBP that were targeted for improvement solely by being informed they were not observed implementing an EBP. Educators in the control group were most likely to change their practices at Time 2 around working one-on-one with children and talking about feelings (100% improved between Time 1 and Time 2) and were least likely to change practices to incorporate math activities (0% improved) or science concepts (15% improved.) (Table 3).

Overall Changes in The Eight Targeted Ebp and Beqi Scores from Time 1 to Time 2

When looking across all educators with complete observational data in both treatment and control groups, however, there were no statistically significant changes in the eight practices between Time 1 and Time 2, although overall means declined in both groups, from 4.74 to 4.56 in the control group and from 5.17 to 4.69 in the treatment group. Additional analyses on the participants whose scores improved from Time 1 to Time 2 indicated that participants with the least amount of PD in the past year, who also had lower scores at Time 1 (r = − .71, p < .001), were more likely to improve rather than decline over the 12 weeks, regardless of membership in the treatment or control group. Educators with more experience were marginally less likely to improve their practices over the 12 weeks (r = − .31, p < .10). There were no changes in overall BEQI scores on target practices from Time 1 to Time 2. (Table 2).

Self-reported Adherence to Goal Setting

At weekly intervals ten times over the intervention period, we asked educators to report on the number of days per week they had implemented their goal. Using a repeated measures ANOVA (F (1, 22) = 620.21, p < .001), we found statistically significant differences in the number of reported goal days over the course of the intervention. Reported goal days were lowest towards the start of the intervention and were greatest during the mid-point of the intervention for Goal 1 and at the end of the intervention for Goal 2. However, we did not find an association between the number of days educators reported achieving their goals and observed practices at Time 2. See Fig. 5.

Fig. 5
figure 5

Self-Reported Goal Days over the course of the interventionsss

Evaluation of Program

Treatment group participants were asked to provide an evaluation of the program after completing the final survey. We asked whether they felt that the program helped them better support the children in their program, and whether they would recommend it to other educators. Ninety-six percent thought it was helpful to receive immediate and strengths-based observation feedback; 91% thought it was helpful to reflect weekly on their practices; 73% would recommend the BEQI goalsetting and text message nudge intervention to other educators; and 70% believed it was beneficial for the children in their care.

Discussion

This study offers initial experimental evidence that a tech-based intervention integrating virtual observation with goal setting and text-message nudges on specific EBP practices that promote child development can be effective. Combining an observational tool that focuses on specific EBP with 12 weeks of feedback, follow up, and support for improvement, our results demonstrate that goal setting and follow up text messaging can encourage the adoption of EBP in childcare settings. We found strong evidence that goal setting and follow-up text messaging is more effective in encouraging adoption of specific practices than observation and feedback alone. This study confirms results from Hanno (2023) demonstrating that text-message nudges may be one important element of scalable, impactful PD for childcare educators. Our study also differs from Hanno (2023) in that the list of potential goals to choose from were identified by researchers rather than by educators. Future work should compare the degree of buy-in for behavior change when goals are selected by educators themselves.

Most importantly, we found effect sizes that are only slightly smaller than effect sizes noted for more intensive interventions lasting for months and/or requiring several hours of training for educators (e.g., Werner et al., 2016; effect sizes ranging from 0.35 to 0.44 for effects on classrooms and caregivers). From this perspective, our results bode well for adding low-cost, easily scalable interventions to the mix of supports available for childcare educators. This may especially be the case for educators who are new to the field, or in low-resource or other types of settings where PD for early childhood educators is not readily available, as we found that changes were greatest among educators with the least amount of experience and training. Given the present shortage of qualified childcare educators, identifying how to begin engaging these educators in PD is an important task. Interventions such as the one we described here thus may provide an important first step towards sustained quality, especially given the positive feedback we received from treatment group participants about their perceived value of our intervention.

However, we also found that observation of EBP at Time 1 did not guarantee continued observation of the same EBP at Time 2. Like conclusions drawn by Hanno (2023), text messaging alone may not be sufficient for sustained reliance on EBP. In our trial, even when educators were affirmed for their reliance on EBP at Time 1, we did not consistently observe the same practices 12 weeks later. Clear focus and goal setting may be essential for maintaining practices over time. We also were able to document educators’ self-reported achievement of goals over the 12 weeks, which demonstrated the great variability in adherence to goals over the intervention period. While this study does not provide insight into why goal adherence went up and down, understanding what helps educators integrate new goals into routine practices is a central question for more deeply understanding how and why this type of intervention might work.

There were several limitations to this study, notably the small sample size and the subsequent inability to adequately test moderating effects. More research with larger samples is needed to build on the findings reported here. Further, we did not repeat the initial needs assessment survey at the end of the study, so we were not able to examine a comparison of Time 1 and Time 2 in all variables across the sample, which may have precluded our ability to detect changes in self-efficacy, ratings of the importance of specific activities and stress levels, as well as the role that knowledge and beliefs may have played in influencing adoption of EBP.

However, even with these limitations, the results from this study point to several conclusions to inform future work in this area. First, observations followed by goal setting and text messaging around specific practices is a viable route towards promoting adoption of quality practices, particularly for educators who have not had much PD and less experience in early childhood. Because many quality improvement programs offer more support to educators with higher quality levels and deeper investment in the field of early childhood, this type of intervention helps fill the gap for educators who are new to early childhood, who have invested less in PD, and/or have less opportunities for or access to PD. Second, experimentation with tech-enabled approaches to quality measurement and improvement is critical to ensure that the early childhood field is maximizing the potential of innovations.

In conclusion, this study breaks new ground by using virtual quality measures to focus on clearly defined, specific EBP, followed by nudge theory to encourage behavior change among childcare professionals. Given the millions of young children in childcare and the shortage of high-quality settings, the field of early childhood is ripe for innovations in training and professional development, especially those that rely on virtual, tech-based approaches that are inexpensive and scalable. Our study demonstrates that using virtual observations and text messaging to deliver nudges holds promise as an impactful and cost-effective intervention to promote quality. Future work should aim to replicate and expand this study, as a way of improving access to quality childcare for all children.