Background

The COVID-19 pandemic has laid bare the costs of failing to access and use scientific research that informs governmental action and communication [1,2,3]. This has led to widespread calls to evaluate effective science communication and dissemination interventions [4, 5]. Government officials have played a particularly meaningful role in communicating about the pandemic via social media, which has been critical for disseminating accurate information but has, in some cases, contributed to the spread of misinformation [4]. Despite existing strategic models of science communication, much of the previous work in this area has been retrospective and relied on self-report of how policymakers access and use research. Few experimental studies have investigated effective strategies to support policymakers’ use of research evidence (URE). This study seeks to prospectively and experimentally evaluate an enhanced science communication intervention to support US legislators’ use of public health research—to our knowledge, the first of its kind [6]. Specifically, we investigate legislators’ URE via social media, given its role in both combatting and perpetuating misinformation during the pandemic.

Study context: the COVID-19 pandemic, misinformation, and politics

In 2020, the world was gripped by the COVID-19 pandemic, which not only disrupted the normal social order (e.g., “work from home”), but also altered the public’s degree of exposure to and engagement with scientific research. The intense need for supporting the rapid uptake of high-quality research evidence encouraged responses from the scientific community. Even before the pandemic, there were significant concerns among the scientific community that a “war on science” was politicizing and delegitimizing public perception of scientific credibility [7]. Despite this popular narrative [8, 9], US and international public opinion of science remained positive among the general public [10, 11] and the policy community [12], plausibly because the “war” was spurred by the beliefs of vocal individuals concerned about vaccines and other specific issues [10]. Immediately in the wake of the pandemic, public trust in science fluctuated [2] and declined over time, especially among Republicans [13]. Despite these recent trends, substantial research has shown that most policymakers across the aisle have reported that research is valuable to their work, and are eager to use research in their work even if approaching it in different ways [14,15,16,17]. Lawmakers also report difficulty accessing unbiased, nonpartisan, or agenda-neutral scientific evidence [18, 19] and view university-based research as more trustworthy than that from advocacy groups and think tanks [20]. This is evidenced in studies with state legislators, including one analysis indicating over half of state legislators as highly valuing research evidence, but less than a quarter reporting access to high-quality research evidence [12].

Mutual mistrust and a lack of scaffolded opportunities to facilitate interactions between academic researchers and policymakers are prominent barriers to URE [21, 22]. It is critical that public health practitioners recognize that trust in science is built through interpersonal relationships. Misunderstanding the need for relational approaches to research translation have inspired simplistic, one-way dissemination efforts that “push” information from research organizations without regard to policymakers’ current needs [23]. Such impersonal science communication approaches are ultimately inconsistent with best practice—which should involve collaborative and interactive approaches for tailoring responses to policymakers’ goals and evidence needs [21]. In particular, research that is timely and relevant is most likely to be used by policymakers [24].

Prior dissemination research

A meaningful body of research has shed light on conditions that foster URE in the policymaking process [21, 25,26,27], yet few studies have experimentally investigated best practices for disseminating research evidence among policymakers, and virtually none have examined the impact of research dissemination on policymakers’ public discourse. Meaningful experimental work has been done recently to improve the way that research is communicated to policymakers. For instance, evoking emotion and threat-based language [28], state-tailored economic evidence [29], and cueing relevance to the state or legislator [24] have been shown to increase legislators’ access of research evidence via email. However, it is unclear what measurable impact such disseminations could have on policymaker URE.

Previous experimental studies with legislative subjects have investigated from whom policymakers learn and how access to information can influence their behavior [23]. For instance, a recent study found that participation in a bipartisan group wherein policymakers discussed their proposed bills with one another increased support for those bills [30]. A subsequent research dissemination study found that when provided with a non-partisan technical briefing on a bill, legislators were 60% more likely to support that bill [31]. Such studies provide evidence that communication can affect policymakers’ behavior [23, 30, 31]. We seek to build on this literature base by further shedding light on how disseminated research information may affect legislators’ URE in their public discourse, specifically in their social media posts.

Conceptual framework

This work draws upon multiple theoretical frameworks for understanding how, why, and when policymakers use research, as well as best practices for disseminating research evidence as revealed in extant literature. Foremost, corresponding with Brownson’s Model of Dissemination Research [32], the current study source involves researchers who send messages about research synthesis in fact sheets, using emails as the channel to deliver the message to an audience of US state legislators and their staff. Embedded in this investigation is investigators’ recognition of a widely used typology describing how policymakers use evidence [30, 31], which includes evidence that directly informs policy development (i.e., instrumental use) or how policymakers think about causes and consequences of problems (i.e., conceptual use). Policies may also leverage research or evaluation methods (i.e., process use) or research may be used to justify preconceived policy stances (i.e., tactical use [19, 20]).

John Kingdon’s Three Streams Model is fundamental to understanding when research evidence is deemed timely based on socio-political factors [35]. Timeliness and relevance of research correspond with these opportunity windows, which has been emphasized for policymakers’ URE [36, 37]. To do this, science dissemination efforts must focus on end-users’ needs, which is exemplified by an array of practices for bridging research and policy (e.g., Family Impact Seminars; Pew Results First, [14, 22]). In contrast to dissemination efforts involving a one-sided flow of information to policymakers, starting with policymakers’ interest areas allows the research translation effort to be targeted in responding with evidence related to current policy windows (i.e., discrete opportunities for policy change [33, 39].

An emerging field studying URE sheds light on why policymakers URE and emphasizes both access to relevant research and the building of trusting interpersonal connections with scientists [21]. Some have deemed dissemination efforts focused on access to be a “first-generation” approach, which, typically share research products not intentionally chosen for their policy relevance. In contrast, “second-generation” approaches involve facilitating researcher-policymaker relationships [21]; these more potent, interactive partnership approaches can be cost-prohibitive because of the staff time required to develop relationships. Fortunately, these approaches are not necessarily mutually exclusive; insights gleaned from interacting with policymakers (via second-generation approaches) may inform and strengthen the disseminations of relevant research evidence, extending the reach and impact of the interactive models. As conceptualized here, an enhanced dissemination (described below) addresses critical concerns about standard dissemination practice by providing relevant and timely research evidence, which is informed by researcher-policymaker interactions, and ultimately increases opportunities for further researcher-policymaker interactions, which then inform content development for future dissemination.

SciComm Optimizer for Policy Engagement (SCOPE) intervention

The SciComm Optimizer for Policy Engagement (SCOPE) is a replicable model for disseminating and improving the reach of research among policymakers. The following core principles of this enhanced dissemination model are intended to address common flaws critiqued in URE literature on dissemination methods [21, 23] by drawing upon theories and prior research on the best practices for bridging research and policy.

Timely and relevant: Policymakers’ perception of relevance may be improved by recognizing their needs and priorities, cueing individual or local relevance, responding with corresponding information, and messaging accordingly [24, 37, 40]. “Feedback loops” can be created by drawing on interactive forms of research dissemination (e.g., partnerships, conversations) to inform relevant content for broad-based dissemination efforts. In particular, this work builds from an experimentally tested model for supporting policymakers’ URE known as the Research-to-Policy Collaboration (RPC [15,16,17]) which facilitates nonpartisan responses to policymakers’ evidence needs by assessing their policy goals before matching them with corresponding researchers. When multiple policy staff have questions about a similar topic, that topic is determined to be politically timely and associated resources are apt for dissemination. We sought to supplement the impact of this interactive brokerage model by expanding the evidence syntheses resulting from researcher-policymaker collaborations. Researchers participating in the RPC rapid response network directly contributed to the written products that were disseminated; thus, SCOPE provided a platform for researchers to communicate directly with policymakers about issues deemed timely by partnering legislative staff. Legislative staff often identified priority areas that paralleled their pre-pandemic interests, but sought to further understand the implications of the pandemic (e.g., on overdose, violence, human trafficking, child abuse).

Researcher-policymaker interactions: SCOPE directly connects researchers and policymakers via email. Researchers develop fact sheets and email them to legislative officials on behalf of authors (i.e., sender is the researcher’s name and message body is a plain text, polite message). This allows the legislative officials to reply directly to the author, who was offered technical assistance (e.g., how to avoid partisan language) and logistic support (e.g., scheduling meetings). Each dissemination prompted, on average, two meaningful interactions (e.g., request to meet; answer questions; present or even testify at a hearing) between the author and a policymaker or staff. Researcher-policymaker interactions informed future disseminations in a feedback loop, which meant that the intervention was not merely a one-way “push” of information, and instead was both responsive to and facilitated interactions between researchers and policymakers.

Brief, skimmable formats: Policymakers have competing priorities and are constantly inundated with a barrage of information ranging across policy issue areas. This results in information overload [41]. In fact, most state policymakers have previously reported they are less likely to read “full reports” than skim information [42]. Moreover, content of research materials should be written in accessible, jargon-free, plain language [37, 43, 44]. Additionally, briefs should be succinct and focus on a singular key point that is immediately conveyed to the reader [32, 44]. Following these guidelines, research-based materials were provided in a brief format preferred by policymakers, such as 1–2 page fact sheets, briefs, or notes. These synthesized implications were based on a body of research instead of single studies and often provided examples of relevant practice.

Continuous quality improvement (CQI) involves cycles of planning, testing, and refining practice strategies to improve practices over time [45]. SCOPE uses a rapid-cycle evaluation method that deploys A/B field experiments to learn what approach to science communication increased the reach and visibility of disseminated fact sheets (e.g., [15, 30]). These CQI efforts serve to improve the reach of disseminations over time by informing an evolution in science communication practice.

Study aims

To experimentally evaluate the SCOPE intervention, a randomized controlled trial was undertaken with legislative offices (Fig. 1). This experiment sought to evaluate the effect of researcher-policymaker engagement on state legislators’ research use in social media posts during the pandemic. We hypothesized that legislators randomly assigned to receive the SCOPE intervention would be more likely to use evidence language in their public discourse related to the pandemic. Investigators felt that a study investigating ways to increase public officials’ use of scientific information in their constituent-facing communications was particularly timely in light of rampant misinformation that became prevalent in public discourse during the pandemic.

Fig. 1
figure 1

Science Communication Optimizer for Policy Engagement (SCOPE) model steps. The SCOPE model builds off of interactive models of research translation to ensure the content disseminated is relevant and timely for officials. Legislative officials who sit on relevant committees receive research syntheses directly from researchers. The content is disseminated through a platform that allows for continuous quality improvement (CQI) via field experiments that inform how to increase the reach of content over time. Thus, legislative recipients are randomized for message testing prior to dissemination and findings from those analyses inform future dissemination efforts. Direct contact between researchers and legislators fosters additional interactions that create a feedback loop in this process, leading to the development of additional content relevant to current policy priority areas

Methods

Intervention delivery

To experimentally evaluate the SCOPE intervention, a randomized controlled trial was undertaken with state legislative offices (Fig. 1). Quorum (quorum.us), a public affairs database with a client-relationship manager, was used to email selected state legislative offices and collect and manage the outcome data. Recipients in the intervention group (state legislators and their staffers, if any) were sent up to 75 emails, roughly one or two per week, containing accessible summaries of research evidence between March 1, 2020, and March 31, 2021. Evidence syntheses were relevant to a range of social issues that were affected by or salient during the pandemic (e.g., substance use, violence, workforce) in addition to public health information (e.g., mask wearing). Corresponding with substantive evidence regarding how policymakers use research, the resources disseminated were timely and relevant and brief and promoted interactions with researchers [25, 38, 45]. Recipients had the option to be removed from receiving future emails. The email body was often brief with four main parts: an introduction of the author/sender, a note on why the resource was relevant at that time, the link, and an offer to answer any questions (samples provided in Supplemental Materials). Emails were sent during normal business hours Monday through Friday. The author/sender interacted genuinely with any responses received, such as answering questions asked via email and participating in meetings. They were provided logistical support for scheduling meetings, a trained facilitator for such meetings, and suggestions on responses to questions if requested. The control group did not receive emails or any resulting opportunity to engage directly with the author/sender.

Sample

US state and territory legislators were chosen based on committee assignments. All US states and territories were included. Specifically, US state legislators who sat on committees related to health were selected for the sample. Randomization was conducted using a simple randomization procedure in Excel by a study team member to sort 25% into the control group and 75% into the intervention. These state legislative offices were randomized such that three-quarters received the intervention (n = 3034) to enhance the overall availability of critical information to decision-makers during the worldwide crisis (control group n = 1016; see Supplemental Fig. 1 CONSORT diagram). Given the magnitude of the crisis, it seemed unethical to withhold information from more legislators than estimated for statistical power purposes. As state legislators vary in individual capacity to check emails [46, 47], the staffers of legislators in the intervention group were also included as recipients. Sample demographic data are provided in Supplemental Table 1. All present data are observational in nature; recipients were not contacted for data collection purposes for the present study.

Measures

The dependent variable of this study is URE in COVID-19-related social media posts as indicated by Boolean phrases detecting linguistic markers for both URE and the subject matter (Supplemental Table 2). These were developed a priori based on a validated coding scheme developed in prior work [39]. Original posts (i.e., no retweets) on legislators’ official Facebook, Twitter, Medium, and YouTube accounts were searched.

URE: In previous work, linguistic markers have shown both to correlate with in-depth, validated coding protocols as well as demonstrate sensitivity and predictive validity for detecting change in policymakers’ URE over time [25, 39]. The study team drew from keyword phrases used in prior studies (e.g., data shows, evidence-based, empirical research, representative sample, conduct a study [20, 21, 41]). This initial list was revised by interviewing coders trained and experienced on a validated, in-depth, deductive coding protocol that was used to quantify URE in hundreds of bills. Coders indicated which keywords had appeared in actual legislative products to denote URE. Additionally, coders were asked how, rather than just if, these keywords are used in bills to create conceptually meaningful subdomains of URE. For example, research can be used to denote the prevalence of a problem (e.g., studies show that 1 in 4 children) or as part of a solution (e.g., develop and implement an evidence-based program), or to hold accountable the relevant parties (e.g., assess the impact of the different programs and produce a report for Congress), among other goals. While all those categories represent more overt expressions of research results or processes, we also captured indirect references to research through a “conceptual” research use category. Research can be used conceptually by alluding to research-based concepts related to health, such as disparities, risk factors, and social determinants, as described in disseminated content.

Throughout this process, each keyword phrase and group were reviewed to assess the prevalence of false positives (e.g., “evidence” referring to forensic evidence required more complex search phrases to assess “research evidence” or “scientific evidence”). Ultimately, eight keyword categories were created and comprised of approximately 200 research phrases. The phrases allowed for examination of a larger amount of data than what could be coded manually, given the large present sample.

Subject matter: Three phrases indicating relevance to the COVID-19 pandemic were identified by the study team a priori to data collection and analysis.

Professionalism: The Squire Index [49] was used to quantify the degree of professionalism of the members’ state legislature. The Squire Index assesses state legislatures’ capacity to create and use information for their policymaking and actively participate in policymaking. It compiles data such as number of staff, legislator salaries, and number of days in session.

Data collection

The number of social media posts containing both subject matter and URE markers posted between March 1, 2020, and June 30, 2021 (to align with the beginning of the intervention and the end of many state legislative sessions, respectively), was used as the dependent variable for analyses. Each URE category phrase was searched within COVID-related social media posts. The counts of how many social media posts a state legislator produced that satisfied these Boolean phrases were returned using Quorum. Re-posts and shared posts (e.g., retweets) were excluded. The counts across URE categories were added to obtain a total count of social media posts related to COVID using URE. Additional data were collected during a baseline period (March 1, 2019–February 29, 2020) to test the equivalence of groups a priori by assessing the number of health-related social media posts made by state legislators in our sample.

Analyses

A zero-inflated negative binomial regression for count data was conducted to investigate differences between the control and intervention groups’ URE within COVID-related social media posts from March 1, 2020, to June 30, 2021. The distribution of relevant social media posts was skewed positive and met the assumption of overdispersion for negative binomial regressions. Subsequent to detecting a main effect, similar secondary analyses were conducted to examine the ways that legislators used research language in social media posts. Given the distribution pattern of the outcome variable, a low interclass correlation between legislators within states, and optimal model fit indices across potential modeling options, a zero-inflated negative binomial model was selected. The rate of having no social media posts containing both URE and COVID-19 markers in this sample was 69.95%. Excessive zeros contributing to overdispersion of the data may indicate that a portion of state legislators’ social media posts were not captured or that the state legislators do not post to a public social media account. Therefore, we adjusted for the zero-inflation using two indicators that would be likely to predict whether a state legislator created a social media post with URE at all or not—the Squire Rank Index and the intervention. The Squire Index, a measure of state legislative professionalism, indicates the institutional capacity of legislatures for engagement. The Rank version orders the states based on their raw index score [49]. This measure is related to legislative capacity for creating new bills [50]. We hypothesized that capacity for engagement may influence social media activity levels, with members of more professional state legislatures being more likely to be expected to post on social media and have sufficient professional resources to be able to do so. For example, legislators in less professional legislatures are expected to hold jobs outside of their part-time legislative work and therefore may not have personal or staff time to post to social media. In contrast, legislators of more professional legislatures work full-time as a policymaker and therefore have the capacity to post to social media. Investigators accounted for the intervention in both the inflation and count parts of the model because we expected that the intervention could affect both whether someone posts URE at all as well as how often they do so.

An additional set of analyses were conducted comparing health-related social media posts with URE markers during the baseline time period in order to examine differences between groups prior to the intervention.

Results

Descriptive statistics are available in Table 1. No significant group differences were found in the number of health-related social media posts using research language during the baseline time period. Results of the zero-inflated negative binomial regressions examining social media posts during the intervention period are presented in Table 2. Results showed that state legislative offices receiving the enhanced dissemination emails posted 24% more social media messages containing evidence language compared to control offices (IRR = 1.24, p = 0.02; Fig. 2). This effect size translates to an additional estimated 789 posts that used research evidence language across the sample attributable to the SCOPE intervention.

Fig. 2
figure 2

Main effect of intervention on URE in social media. State legislators receiving the intervention from the RPC produced 24% more COVID-related social media posts with URE than those not in the intervention group, controlling for the effect of state professionalism on zero-inflation

Table 1 Descriptive statistics
Table 2 Intervention effect modeled with zero-inflated negative binomial regressions

The Squire Index was included in every model conducted and was a significant predictor of zero inflation for all models (p < 0.001). This indicates that a state’s professionalism is associated with inflated zeros such that legislators in states with low professionalism are more likely to have had zero social media posts related to health that included URE.

Secondary analyses (Fig. 3) were conducted to examine which keyword categories (i.e., problem definition, etc.) were driving this relationship, while controlling for the effect of state professionalism on zero inflation. Legislators who received the intervention produced 67% more COVID-related social media posts containing data/analytics research language than legislators in the control group (IRR = 1.67, p = 0.004), and 28% more posts with conceptual research use (IRR = 1.28, p = 0.01). Additionally, legislators in the control group produced 69% more posts with research use related to knowledge generation (IRR = 0.31, p = 0.002) and were 68% more likely to post research use related to knowledge generation at all (OR = 0.32, p = 0.04) than those in the intervention group. Analyses for Accountability and Type of Study categories, which were created prior to data collection and not otherwise examined until analyses, yielded large confidence intervals attributable to a very low, restricted range of use of these indicators; therefore, investigators did not interpret these coefficients. There were no significant differences between groups for other keyword categories (problem definition, problem solution, methods).

Fig. 3
figure 3

Secondary analyses: intervention effect on URE categories in social media. The relationship between the intervention and URE in social media products was driven primarily by the Data/analytics and Conceptual sub-categories. State legislators in the intervention group produced 67% more COVID-related social media posts with URE falling in the Data/analytics category than those not receiving the intervention, controlling for the effect of state professionalism on inflation. Legislators in the intervention group produced 28% more COVID-related social media posts with URE in the Conceptual category than those not receiving the intervention, controlling for the effect of state professionalism on inflation. Legislators in the control group produced 69% more social media posts with URE in the Knowledge Generation category

Discussion

This study suggests that it is possible for theory-based, enhanced research dissemination efforts to influence state legislators’ public discourse by increasing their access to scientific evidence. This is noteworthy because science-based discourse may counteract misinformation and support the government’s role in sharing accurate health information with the public [4]. In particular, SCOPE was most effective for improving conceptual URE via reference to theoretical and research-based language, as well as data and analytic language that may reflect a type of process URE. Intervention legislators may have been more likely to emphasize high-quality methods indicating credibility behind scientific concepts. These findings correspond with research suggesting that policymakers often use research indirectly for understanding how to conceptualize problems [31], demonstrating credibility among colleagues, and educating constituents [38].

In contrast, intervention legislators posted less often about creating and disseminating evidence (i.e., knowledge generation). Perhaps intervention offices perceived existing research to be sufficient in supply, in line with the availability heuristic, therefore feeling less need for additional studies [51, 52]. Alternatively, control group legislators may have sought more research to clarify contradictory sources of information. Due to the novelty of this area of research, comparable literature to assist in this interpretation is limited. Further studies are needed to understand the relationship between exposure to research evidence and demand for additional scientific knowledge.

This work builds on literature related to research dissemination and use by addressing some consistent methodological limitations. Previous work on supporting policymakers’ access to and use of scientific information is primarily retrospective, correlational, representative of only a small portion of legislators, and subject to social desirability (i.e., rely on self-report [45]). Very little rigorous research has been done to evaluate research translation strategies [6], and this study demonstrates a feasible method of experimentation of current practices by drawing upon observational measures of impact such as public discourse. Such evaluation approaches could be highly beneficial for well-established and well-respected research synthesis centers (e.g., McMaster Health Forum, Scholar Strategy Network, and synthesis centers [53]). Not only did investigators prospectively examine the impact of a research dissemination model with experimental design, but also this study incorporated leading theories from this past research directly into the design of the present intervention. Thus, this rigorous evaluation provides some causal evidence of effectiveness for theory-based research dissemination practices that have the potential to influence legislative discourse at a time of worldwide crisis.

Limitations and future directions

While this study revealed the potential for SCOPE to significantly impact research language use, given the nascency of this topic and methodology, there are potential limitations that warrant further study. First, the present work compared SCOPE to a control that did not receive evidence in a way we can track; there is more work to be done to compare this theory-based model to one that follows the first-generation “push” approach to dissemination.

Further, the estimates of the effect may be an underestimate, given that legislators within a state legislature are inherently connected—sharing buildings, committees, and often even staffers—which may contribute to a spillover effect. The social use of evidence for earning trust, demonstrating credibility, educating others, and enhancing debate and collaboration between lawmakers has been documented in previous studies [18]. Therefore, the control group may have been exposed to the intervention vicariously through the intervention group. The effects may be underestimated further given the social media data were restricted to posts specific to the pandemic (as opposed to broad social issues mentioned in briefs); this provided a feasible amount of data for processing capacity of the data collection platform. Additionally, due to large sample sizes involving copious amounts of analyzed text, assessing the nature of the content or veracity of research used in social media posts was beyond the scope of this study. It was not possible to do in-depth coding regarding the nature of how evidence was used in posts; thus, the current study examined overt indicators of evidence language. Future work may seek to incorporate measures of content accuracy and types of research use in order to obtain a more comprehensive understanding of policymakers’ use of high-quality scientific evidence.

Finally, it is unclear how the current pandemic context affected demand, access, and uptake of research information; therefore, the findings are not generalizable to other time periods or other topic areas (e.g., climate change) and there is a need to replicate this experiment. Converse to generalizability across issue areas is a need to differentiate strategies for different types of policymakers. For instance, while state legislators are critical for public communication and enacting new policies, agency officials may be another audience of interest for future intervention testing given their role in implementing policy. Similarly, research use patterns may vary by political affiliation as prior studies have illustrated different research use patterns, emphases, and impact of research communication approaches across party lines [16, 17, 29]. Future studies should consider experimentally testing strategies tailored to sociopolitical preferences and testing interventions by stratifying samples by political party.

SCOPE could be replicated in future science translation efforts that embed both the relational nature of the enhanced dissemination method and routinized evaluation for continuous quality improvement. To improve access to trusted, nonpartisan research evidence and to facilitate its use, research translation teams must embrace best practices such as personalizing to the needs of policymakers. Those seeking to apply these findings should adhere to the relational aspect of this work, such that the science that is distributed is drawn from interacting with policymakers, and the dissemination effort itself encourages further interactions between researchers and policymakers. More broadly, these findings could be built upon by further increasing the reach of the research syntheses in this effort, such as using different interpersonal tactics or framing strategies [28, 55, 56]. Moreover, expanding the interactive nature of this work with further optimization may have the potential to yield additional or more complex forms of URE.

Conclusion

It is critical to the widespread implementation of evidence-based programs to promote access and use of research evidence among the people who determine the nation’s priorities, write our policies, and deem what gets funded. This work highlights the importance of developing evidence-based strategies to support decision-makers’ research use. Effective research translation approaches have the potential to improve the societal impact of scientific knowledge ranging from battle-tested practices to revolutionary studies as seen in the pandemic. As the COVID-19 pandemic has shown, the spread of both scientific information and inaccurate misinformation is complex and dynamic, which requires multifaceted research dissemination strategies. Rigorous evaluations of research translation efforts like this one are key to developing and identifying evidence-based strategies for improving the use of science.