Skip to main content

Ensuring Prevention Science Research is Synthesis-Ready for Immediate and Lasting Scientific Impact


When seeking to inform and improve prevention efforts and policy, it is important to be able to robustly synthesize all available evidence. But evidence sources are often large and heterogeneous, so understanding what works, for whom, and in what contexts can only be achieved through a systematic and comprehensive synthesis of evidence. Many barriers impede comprehensive evidence synthesis, which leads to uncertainty about the generalizability of intervention effectiveness, including inaccurate titles/abstracts/keywords terminology (hampering literature search efforts), ambiguous reporting of study methods (resulting in inaccurate assessments of study rigor), and poorly reported participant characteristics, outcomes, and key variables (obstructing the calculation of an overall effect or the examination of effect modifiers). To address these issues and improve the reach of primary studies through their inclusion in evidence syntheses, we provide a set of practical guidelines to help prevention scientists prepare synthesis-ready research. We use a recent mindfulness trial as an empirical example to ground the discussion and demonstrate ways to ensure the following: (1) primary studies are discoverable; (2) the types of data needed for synthesis are present; and (3) these data are readily synthesizable. We highlight several tools and practices that can aid authors in these efforts, such as using a data-driven approach for crafting titles, abstracts, and keywords or by creating a repository for each project to host all study-related data files. We also provide step-by-step guidance and software suggestions for standardizing data design and public archiving to facilitate synthesis-ready research.

Systematic reviews are an important tool for informing policy and practice. When done well, they provide a robust synthesis of available evidence, account for risks of bias in individual studies, and address problems associated with the volume of primary research being published (Johnson & Hennessy, 2019; Wilson & Tanner-Smith, 2014).

To conduct a comprehensive evidence synthesis, such as a systematic review and meta-analysis, primary studies related to the research aim must be identifiable and report relevant study and outcome data. Those who conduct evidence syntheses — hereafter referred to as evidence synthesists — often need information about study processes and measured variables aside from a manuscripts’ primary outcome of interest to explore why there are differences (i.e., statistical heterogeneity) between studies examining the same intervention. This information is often missing, which leads to uncertainty about the generalizability of intervention effectiveness across different groups (e.g., race/ethnicity, gender, age) and can make these efforts slower and less accurate (Borah et al., 2017; Haddaway & Westgate, 2019; Nakagawa et al., 2020; Nuijten et al., 2016). Even when studies report their statistical data and results in detail, there is often still important information missing about other factors that require further attention before they can be accurately synthesized (e.g., description of the intervention, population, or outcome measures). Data and meta-data may be missing from a primary study for many reasons, including article length restrictions, a lack of detailed reporting standards or awareness of the importance of open data, or a desire to retain control of analyses, amongst others.

Given the potential impact of their work on human health and behaviors, prevention scientists should be at the forefront of ensuring primary research is transparent, reproducible, and furthermore, we argue, synthesis-ready. Synthesis-ready research is our term to describe a study (e.g., randomized controlled trial (RCT), quasi-experimental study (QED), or observational study), its meta-data (i.e., descriptive information about the study’s data, including protocols, tools, algorithms, so that one could replicate the study), and data (information collected while conducting the study) that can be readily and accurately used in evidence syntheses.

Resources are available to enable authors to share detailed data and meta-data associated with a study; yet, these resources are often not directly connected to scientific journal practices. Some of these options make it much easier to discover, use, and cite authors’ research, substantially increasing the impact and legacy of their work. Thus, our aim was to draw on resources from across disciplines (Grant et al., 2018; Montgomery et al., 2018; Schulz et al., 2010; Wilkinson et al., 2016) to produce a guide to reporting primary research that is open, accessible, and synthesis-ready. To illustrate synthesis-ready research principles and streamline this process for prevention scientists, we briefly present an empirical example from a mindfulness meditation RCT to improve college student well-being (“Introduction to Empirical Example”), followed by a demonstration of the application of best practices for synthesis-ready research to this example (“Evidence Synthesis Process Applied to Our Empirical Example” and “How to Make Research Synthesis-Ready: Steps from Study Design to Completion”).

From Primary Study Research to Evidence Synthesis

Primary study conduct and reporting guidelines now advocate for reporting enough information for research to be transparent and reproducible and also provide specific information on how to do so (Chan et al., 2013; Grant et al., 2018; Schulz et al., 2010). Similarly, efforts have focused on improving the degree of FAIRness of primary studies, by making research Findable, Accessible, Interoperable, and Reusable (Wilkinson et al., 2016).Footnote 1

Findable indicates research is identifiable to anyone searching for it; data and meta-data and their locations clearly described. Data and meta-data that are retrievable by their identifier (e.g., a digital object identifier, DOI) and are freely obtainable in full without barriers to access (e.g., without needing to register with an external website or contacting study authors) are considered Accessible. Data and meta-data that are understandable across humans or machines without the need for specialized tools or translation and include linkage to their relevant data/meta-data are Interoperable. Finally, Resuable refers to the degree to which information can be appropriately repurposed (e.g., integrated/built upon). Thus, research data and meta-data should have a data usage license, detailed records of data origins, and be clearly and richly described in their accompanying materials. By following Open Science recommendations and addressing FAIR principles (Frankenhuis & Nettle, 2018; Vicente-Saez & Martinez-Fuentes, 2018), prevention scientists can provide synthesis-ready research that improves the translation of research into practice.

Evidence syntheses rely on the data reported in research studies and by the researchers themselves. Similar to the conduct of primary research, there are best practices for conducting and reporting of evidence syntheses, including the following: problem formulation, systematically identifying and selecting studies for inclusion, coding studies for key features, calculating effect sizes (if synthesizing quantitative data, such as by conducting a meta-analysis), analyzing the data (assess quality and risk of bias, synthesize coded information from studies), and interpreting and communicating results (Online, Fig. 1). The content and the comprehensiveness of the review (if the review authors follow best practice guidelines) are a direct result of the available information reported in primary studies.

Potential Advantages of Synthesis-Ready Research

Synthesis-ready research can increase the transparency, integrity, and reproducibility of research, ultimately leading to improved scientific evidence (Beugelsdijk et al., 2020). Pre-registration and subsequent data sharing practices should reduce the prevalence of post hoc hypothesis (HARKing; Forstmeier et al., 2017; Kerr, 1998) and falsifying or misrepresenting data in results (Banks et al., 2019; Miyakawa, 2020). Doing so can reduce duplication of effort and introduce findings to the research community sooner and ensure that all findings (beneficial and adverse) are reported, results that can prevent costly mistakes and accelerate the impact of prevention scientists (Gupta et al., 2015; Pasquetto et al., 2019). Indeed, synthesis-ready data can be used repeatedly and for multiple purposes, ultimately informing more research questions: thus it is more cost-effective when compared to data that is only used by the team that collected it.

Primary researchers who make their work synthesis-ready should also benefit from these efforts. Synthesis-ready research is more likely to be identified and used in evidence synthesis (Christensen et al., 2019; Gerstner et al., 2017). For example, a recent study showed that hosting data in a repository resulted in 25% more citations than merely stating data were available by request (Colavizza et al., 2020). Additionally, depending on the organization of the meta-data and the data repository used, the dataset may be cited directly, as may other study materials such as the analytical code, further enhancing the profile of the work and those who conducted it (Crosas, 2013).

Because synthesis-ready research is also more identifiable, there will be increased opportunities for future collaborations (Popkin, 2019), which potentially improves future resource-sharing and reduces costs. Producing synthesis-ready research can also increase the recognition of an individual study team to an international audience. Many funding sources require data sharing of their grantees and necessitate the inclusion of such plans in grant proposals (e.g., Wellcome Trust, the UK Medical Research Council, and the US National Institutes of Health: Gaba et al., 2020; NIH Data Sharing Policy & Guidance, 2003; Wellcome Trust Public Health Research Data Forum, 2012). Thus, researchers who prepare and deliver grant-funded research findings with synthesis-ready considerations are considered responsive to funders while potentially attaining these other discussed benefits.

Introduction to Empirical Example

Our empirical example is a RCT (N = 140 participants; registered in Clinical Trials, NCT03402009) conducted across two university semesters in the 2018 academic year (Acabchuk et al., 2021). The study evaluated two active treatment conditions to determine what tools best assist university students in developing a personal meditation practice to self-manage symptoms of depression, anxiety, and stress. One group was assigned to the 10% Happier meditation phone app (labeled the “App Group”; Ten Percent Happier: Mindfulness Meditation Courses, 2020), which provides guided meditations, and the other was assigned to the 10% Happier meditation phone app in combination with a Muse EEG neurofeedback device (labeled the “Muse Group”; Choose Muse, 2018). The study included a baseline assessment, randomization to one of two groups, an orientation session for both groups, collection of salivary biomarkers, and follow-up survey assessments on a variety of well-being and behavior outcomes throughout the study (Study Timeline: Online, Fig. 2).

Evidence Synthesis Process Applied to Our Empirical Example

Drawing from our empirical example, if one sought to conduct a systematic review of all types of mindfulness interventions for college students on health-related outcomes (Problem Formulation), they would first need to be able to identify all primary research matching their inclusion criteria (Finding and Selecting Studies). Clearly delineating the research problem is a foundational step in a review, as it drives the search strategy and ability to accurately identify and synthesize primary study research. Research questions in an evidence synthesis should address all the main elements of the studies to be included; many draw on the PICOS framework with its focus on the population(s), intervention(s), comparator(s), outcome(s), and study design(s) elements of a research problem (Stern et al., 2014). Once the relevant literature is identified, data are extracted from the studies (Code Studies for Relevant Features) so that the analysis can commence by synthesizing coded data (Analyze the Systematic Review Database).

Within this broader process, the types of data and meta-data that need to be reported in a primary study depend on the task and/or type of evidence synthesis. For example, research methods need to be fully reported to assess risk of bias. If the review aimed to measure the effectiveness of mindfulness interventions on depression symptoms using a meta-analysis, then primary studies should report the data necessary to calculate an effect size from their depressive symptom assessment for all participants (e.g., means, standard deviations, and sample sizes for each group). To address relevant potential effect modifiers (e.g., age, year in school, gender, intervention fidelity), any data collected during the primary study on participant characteristics and the study methods (e.g., description of groups, duration of intervention) needs to be transparently reported. To examine geographic and temporal effect modifiers, key information about the study location and timing should be provided (Johnson et al., 2017). Additionally, if the synthesis focuses on participants’ perspectives on why interventions are more or less effective, they would need access to any qualitative data collected as part of the study (e.g., interviews, focus groups). Given this variability in the type of questions that could be addressed in a synthesis, it is important for investigators to transparently and comprehensively report a variety of information about their research.

How to Make Research Synthesis-Ready: Steps from Study Design to Completion

Approaching every step of a research project with data sharing in mind is useful and often necessary to ensure that data sharing is possible. This approach has the added benefit of encouraging better overall project management, including clearer workflows and comprehensive documentation of the analytical process. We now illustrate each step of these suggestions from study design to sharing with examples from our mindfulness trial (Fig. 1).

Fig. 1
figure 1

Best practices in generating synthesis-ready research applied to mindfulness trial. DAS, data availability statement

Study Design

Preparing research for sharing can be a resource-intensive process, and so planning and provisioning for this task are vital to consider when submitting funding applications and considering funders’ reporting requirements. When conducting research with human participants, advance planning for data sharing is essential because it ensures appropriate documentation is prepared. Any potential ethical concerns with data sharing should be addressed during the review of study materials by the governing Institutional Review Board, which ensures that participant consent documents include the correct language to ensure participants can consent to anonymized data sharing (Ohmann et al., 2017; Thorogood & Knoppers, 2017). The Inter-university Consortium for Political and Social Research (ICPSR) recommends carefully constructing these sections and provides sample text to use to ensure that participants are informed about the potential use of their anonymized data in future research (Inter-university Consortium for Political and Social Research, 2010).

The study protocol is the foundation of the study, detailing all the processes, procedures, and materials planned for a research project. Best practice for primary studies and secondary data analysis is to register a study protocol that includes primary questions and/or hypotheses, methodology, and planned analyses. With careful attention to developing a comprehensive protocol using a standard structure, this document can increase the synthesis-readiness of the research. The information provided in a study protocol includes important meta-data that may not be available in a single published manuscript but that can be used by evidence synthesists to address a variety of research questions, including using it to assess study quality and risk of bias due to selection of reported results or “outcome switching” (Falk Delgado & Falk Delgado, 2017; Kahan & Jairath, 2018). We recommend that authors review published guidelines for the best tool for the study design in question: The EQUATOR network provides an online resource to compare these tools ( Online Tables 1 and 2 present a mapping of critical elements for a study protocol (using the Standard Protocol Items: Recommendations for Interventional Trials (SPIRIT); Chan et al., 2013) and for a study manuscript (using the Consolidated Standards of Reporting Trials for social and psychological interventions (CONSORT); Grant et al., 2018)) and indicate how study information provided following these guidelines could be used in an evidence synthesis. Table 1

The protocol for our mindfulness trial uses the SPIRIT guidelines (Chan et al., 2013) and is archived in a Zenodo repository (DOI:

Study Conduct

To facilitate data sharing and to distribute the workload across the projects’ lifecycle, transparent and reproducible data management workflows should be employed throughout the research process, including (1) defining data file contents; (2) clearly describing variable names, definitions, units, and formats; (3) implementing consistent data organization and file structure; (4) performing basic quality assurance to ensure that values fall within expected ranges; and (5) providing documentation including the analytical steps taken (crosswalks), data dictionaries (i.e., codebooks), and meta-data (Barton et al., 2010; Strasser, 2012; Yenni et al., 2019). Building in these meta-data collection steps into the analysis process makes it easier to create items like variable crosswalks and data dictionaries to accompany shared research.

Study Dissemination

The most labor-intensive stage of making data synthesis-ready within a FAIR perspective occurs during preparation for study dissemination as it involves several steps. To ensure their research is Findable and Accessible, authors should (1) ensure all authors have a unique permanent identifier such as an ORCID iD (Open Researcher and Contributor ID, 2009) and include this identifier with manuscript and data submission (because institutional emails are often impermanent); (2) include a detailed data availability statement with manuscript submission; (3) assign data and meta-data a globally unique and persistent identifier (e.g., DOI), which ensures long-term discoverability; (4) use available data ‘tags’ in the archiving repository; (5) use consistent names/usernames across the different journal and archiving platforms; and (6) craft the title and abstract and select keywords that will optimize the chances of research being identified via the most typical evidence synthesis method of searching for research (i.e., electronic database searches). To ensure their data is also Interoperable and Reusable, researchers need to attend to licensing, data file structures and formats, and detailed meta-data when preparing their data for sharing.

Data Availability Statements to Ensure Findability

A data availability statement (DAS; also known as “data access statements” or “availability of materials”) provides information about how and where other researchers can access the study data and is often required when submitting a manuscript for publication (Federer et al., 2018; Graf et al., 2020). In some closed-access journals, DAS are treated as article meta-data, so they are accessible even if the article is behind a paywall, ensuring the data are findable. Thus, the DAS should be as explicit as possible, even if some information that is in the manuscript is repeated in the DAS.

The DAS for our mindfulness trial example states: “Both the full raw dataset and the finalized analytical dataset used in this analysis are available from Zenodo (DOI:, along with comprehensive data dictionaries and crosswalks.”

Creating Titles, Abstracts, and Keywords to Optimize Findability

Carefully crafting titles and abstracts, in addition to choosing appropriate keywords, is essential to help other researchers find your research. As much as possible given word limits, these should describe the condition, exposure, and settings of a study (Grant et al., 2018; Montgomery et al., 2018). These pieces of the submission are often the last to be created but are the main pieces of information indexed and displayed in electronic databases and are used to identify relevant research for an evidence synthesis. As a result of their ability to increase the findability of research, we suggest a data-driven approach to crafting these sections. The litsearchr package for the R programming environment offers a systematic approach to generate a list of words likely to improve the identifiability of the study manuscript (Grames et al., 2019; R Core Team, 2019).

Using our mindfulness trial example, we began with a set of keywords we thought likely to capture a set of relevant published articles from which we could draw the most highly used keywords. We conducted a search using these keywords in PubMed on August 24, 2020 and used litsearchr to systematically generate a list of important terms. The results suggest that several terms could be strategically incorporated in the title, abstract, and keywords to increase the findability of the trial (Online, Table 3 provides further details using our example).

Data Licenses

Assigning a license to the data being shared is important to govern how the data can be reused and, more importantly, to help ensure that the study team is credited.Footnote 2

For our mindfulness trial example, we chose the CC-BY-4.0 license, which allows the reuse of the data provided the original researchers are credited with the creation of the dataset.

Data File Structures and Formats

Sharing individual participant data, i.e., data on each variable collected for each participant, affords those wishing to reuse or synthesize these data a range of advantages. Access to data on all variables collected, rather than just those used in the primary study’s analysis, allows synthesists to ensure that an estimate from a primary study is adjusted for a common set of variables when it is included in a meta-analysis, even if the published effect estimate for that study was not (Riley et al., 2010). Even among individual participant data, different levels of granularity exist.

It may not be possible to share each individual participant data point, especially when participant confidentiality is at risk, for example, when studying a rare disease or among participants of specialized programs. In these cases, there are several options for sharing individual participant data. The first is to generate a “synthetic” dataset that mimics the individual-level data by preserving statistical properties and the relationships between variables. Software, for example, the synthpop R package, can help users generate a synthetic dataset (Quintana, 2020, provides a tutorial on using this package). A second option is to remove some individual characteristics that might give away participants’ identities, such as age, sex, and location of recruitment source, but provide the rest of the variables as individual data points. Finally, if it is not possible to share any individual participant data, then at the very minimum, summary-level data (e.g., means and standard deviations or percentages which describe the distribution of all variables between intervention/exposure groups) should be provided for all variables included in the paper and, if possible, the dataset, not solely those thought by the primary study team to be relevant. Of course, for investigators conducting years of longitudinal studies examining numerous questions for which they are not ready to share data, we advise authors to explicitly state in their DAS the reasons for not sharing data. By beginning to capture the true reasons for not sharing data, data availability statements can then begin to be used as support for introducing better protection for and citation of data producers.

To illustrate the distinction in the level of data shared, for our mindfulness trial, we shared the raw data exported from the survey instrument, which contains each participant’s answer to each individual question, and the processed individual participant data, which include scale score summaries (i.e., total DASS-21 score, comprised of 21 individual questions). To explore these different datasets, see the “REDCAP_data.xlsx” and “combined_data.xlsx” files in our Github repository which is archived by Zenodo (DOI:

Study data can be shared as machine-readable (e.g., unformatted Excel/CSV files and raw DNA microarray sequences) or human-readable formats (e.g., formatted PDF tables, figures, and data contained in text such as “There was evidence of an association (Odds ratio: 0.91, 95%CI: 0.85–0.96)”). Although tools are being developed to extract data automatically from human-readable sources for evidence syntheses (Marshall et al., 2020), it is often challenging to import these data into statistical analysis software, reducing interoperability and synthesis-readiness.

For our mindfulness trial example, all study data are shared via Excel files in our Github repository which is archived by Zenodo (DOI: This includes the mindfulness trial’s quantitative survey data. It also includes qualitative data, which were generated from in-person, semi-structured interviews and coded to examine adherence and problems with using the tools. To demonstrate synthesis-readiness, the relevant responses for qualitative coding are included in the data file rather than the full interview notes.


Providing a README, a plain text file that contains useful information about data or software, can help ensure that data are correctly accessed, interpreted, and reused by other researchers. The README file should include a (1) definition of the primary research aims using a relevant alternative framework that addresses the main components of the research which will help others identify and use the research (e.g., PICOTS, TOPICS + M; Haynes, 2006; Johnson & Hennessy, 2019; Stern et al., 2014); (2) a short description of all the data files included in the project repository; (3) signposts to the detailed data dictionaries for each data file which contain definitions of all variables, including coded (e.g., “CRP_t2”) and interpretable (e.g., “C-Reactive Protein levels at 2 months post-baseline”) variable names, the values that each variable is allowed to take, and include missing data indicators and units of measurement; (4) a summary of data processing steps that were used to produce the final analytic datasets from the raw data files; (5) a description of each data file’s relationship to other data files in the repository, and relevant linkage variables, such as participant ID; (6) permanent author contact details; and (7) details of the license, including the full name of the license holder, the year, and where the full text of the license can be found.

The README for our mindfulness trial contains the seven main elements necessary for a comprehensive meta-data file; see Fig. 2 for a snapshot of these pieces and visit Zenodo to view and download all files:

Fig. 2
figure 2

source files, locations, and relationships between the files; d provides permanent author information and licensing information

Mindfulness trial metadata included in the README file. Panels a and b provide the study description according to the PICOTS framework; c provides data

Archiving and Sharing Research Data

Research data sharing can be done via institutional and non-institutional repositories or alongside published articles (Table 1). We have ordered these sharing options from best practice (most likely to be synthesis-ready) to good practice (still synthesis-ready but may require additional effort to prepare for synthesis).

Table 1 Comparison of the options for sharing data

Non-affiliated Repositories

Non-affiliated repositories are commercial or charity repositories that allow researchers to self-archive, and assign a DOI to, their research data. Examples include Dryad, Open Science Framework, and Zenodo (Dryad, 2020; Open Science Framework, 2020; Zenodo—Research Shared, 2020). The main limitation of these repositories is that there are often conditions on how much data can be stored; some charge fees for storage.

For our mindfulness trial example, the repository was created in GitHub and archived using Zenodo, as GitHub deposition is non-permanent and does not assign a DOI to the uploaded data. Repositories of this type can be embargoed prior to publication to prevent “scooping,” and some repositories also provide services to help prevent scooping. For example, projects hosted in the Open Science Framework can have “anonymous” links to remove identifying meta-data as part of double-blind peer-review requirements.

Institutional Repositories

An institutional repository is one hosted by an academic institution, such as the ICPSR (ICSPR, 2021). An advantage of affiliated repositories is that some provide staff to guide researchers who are new to data sharing. These repositories are less well-known, which may limit the findability of data when compared to other options. The terms of deposition should be reviewed to ensure that authors retain data ownership and copyright over the deposited materials.

Supplemental Files

A common option for sharing research data is as a supplementary file to the published manuscript. However, this method has several major limitations. This approach reduces data findability because datasets shared as supplementary material are grouped under the publication’s DOI and are not uniquely identifiable. There is also the potential for supplementary files to become disassociated with their respective publications, for example, during a website redesign. If the published article is behind a paywall, then data attached to it as supplementary files may also be blocked and so data shared as supplemental files may not be fully accessible. Sharing data as supplementary files to an open-access publication should be considered the minimum level of data sharing needed to make research synthesis-ready.

Potential Challenges and Future Considerations

In this article, we have outlined the benefits of synthesis-ready research and provided guidelines on how to follow best practice and produce synthesis-ready research. Yet, we appreciate that there are many perceived and actual challenges to doing so, including costs to data sharing, lack of incentives, lack of awareness or training, not wanting to “give away” data or get “scooped” on findings, and the necessity of ensuring participant confidentiality (Evans, 2016; Tennant et al., 2019; Tenopir et al., 2011, 2015; van Panhuis et al., 2014; Walters, 2020).

To address the lack of knowledge and training, we have provided information, workflows, and a full demonstration using an empirical example, in addition to signposting to further literature, to help researchers plan for the resources needed for making research synthesis-ready at project conceptualization. Many funders have recognized the value of data sharing and will provide funds to do so based on investigator budgets which can offset the tangible (platform sharing fees) and intangible (time spent preparing data) costs of this process.

Researchers have valid concerns about having their research scooped, and yet, several measures can help prevent scooping including pre-registering study analysis plans and depositing preprints while awaiting publication, all core practices of open science. An additional solution is to license data at the point of publication or archival (“How to Make Research Synthesis-Ready: Steps from Study Design to Completion”). Finally, some journals have implemented “scoop-protection” policies for research, a system-level solution we hope more journals will implement to address this challenge (Transpose, 2020). Ensuring that all types of research data are reused appropriately is a warranted concern (Bishop, 2009; Yardley et al., 2014); yet, evidence suggests that individuals reusing data seek to collaborate with the original data creators to harness their knowledge about the data, rather than using it to scoop findings or to disprove the original publication’s findings, resulting in improved and more accurate data reuse and increased opportunities for publication (Pasquetto et al., 2019). As well, providing clearly detailed documentation in the data and meta-data files (“How to Make Research Synthesis-Ready: Steps from Study Design to Completion” and Fig. 2) should reduce inappropriate use of shared data. Finally, we have provided several different options for maintaining the confidentiality of participants while attending to synthesis-readiness (“How to Make Research Synthesis-Ready: Steps from Study Design to Completion”).


Despite the importance of making research available and synthesis-ready, not all scientific disciplines have embraced data/meta-data sharing and transparency at the same rate (Graf et al., 2020; Hardwicke et al., 2020; Sholler et al., 2019; Vasilevsky et al., 2017). If prevention scientists share research in synthesis-ready forms, the field would see improvements, especially in terms of impact on policy and practice. We presented methods for managing data sharing processes that should make it easier to produce synthesis-ready research. Changes in research practice that focus on synthesizability will reduce the delay between implementing effective primary study approaches (Nakagawa et al., 2020) and should have a positive impact on translation of prevention science into policy and practice.


  1. Although FAIR principles primarily refer to enhancing the ability of machines to find and use data, these recommendations also enhance data reuse by individuals. Where appropriate, we have adapted the FAIR components to more directly apply to human-readability efforts for the purposes of this manuscript.

  2. Comparing the full suite of licenses available is beyond the scope of this article; online resources such as ChooseALicence ( and Creative Commons ( provide detailed descriptions of the range of licenses available for data sharing.


Download references


The paper was first developed during the 2019 Evidence Synthesis Hackathon.


The Fenner School of Environment & Society (ANU), the University of New South Wales, and the Faculty of Humanities at the University of Johannesburg funded the Evidence Synthesis Hackathon 2019. EAH has support from NIAAA (K01 AA028536-01). MJP has support from an Australian Research Council Discovery Early Career Researcher Award (DE200101618). LAM has support by an NIHR Doctoral Research Fellowship (DRF-2018–11-ST2-048).

Author information

Authors and Affiliations



Project administration (EAH; LAM), roles/writing — original draft (EAH; LAM), visualization (EAH; LAM); data curation (RLA; LAM); resources (LAM); funding acquisition (NRH); investigation (EAH; LAM; RLA), methodology (EAH; LAM); software (EAH; LAM); conceptualization (All authors), writing — review and editing (all authors).

Corresponding author

Correspondence to Emily A. Hennessy.

Ethics declarations

Conflict of Interest

The authors declare no competing interests.


The views expressed in this article are those of the authors and do not necessarily represent those of the NIAAA, NHS, the NIHR, or the Department of Health and Social Care.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Supplementary Information

Rights and permissions

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Hennessy, E.A., Acabchuk, R.L., Arnold, P.A. et al. Ensuring Prevention Science Research is Synthesis-Ready for Immediate and Lasting Scientific Impact. Prev Sci 23, 809–820 (2022).

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI:


  • Evidence synthesis
  • Meta-data
  • Reporting
  • Reproducibility
  • Synthesis-ready research
  • Transparency