Background

The recruitment phase of multi-site clinical trials represents a large, but modifiable, component of total trial duration and cost [1, 2]. “Successful” recruitment for a trial includes not only accruing the sample on schedule [3] but also enrolling a sample that is representative of the larger disease-suffering population [4,5,6,7]. There are several known barriers to participating in clinical trials, particularly in historically underrepresented communities [8, 9]. Few interventions have been demonstrated to overcome these barriers [10].

The evaluation of recruitment strategies is generally focused on actual enrollment, including successful screening and randomization of study participants. However, measuring activity prior to trial enrollment may provide valuable information for both central and local efforts to accelerate and diversify recruitment, instruct resource expenditures, adjust recruitment campaigns, and, if needed, amend protocols to address observed selection bias.

Efforts to capture prescreening recruitment data may be particularly valuable in preclinical Alzheimer’s disease (AD) trials. These trials enroll cognitively unimpaired older volunteers who are screened for biological markers of AD [11]. Traditional clinical trial recruitment methods may not be effective in these trials, especially when the goal is to recruit a demographically representative cohort. In the first multi-site preclinical AD trial, the Anti-Amyloid treatment in Asymptomatic AD (A4) study [12], it was shown that participants from underrepresented racial and ethnic groups were more frequently recruited through local, compared to centralized or national, efforts [13]. Yet, it has been difficult to determine the most effective recruitment strategies and whether specific central efforts led to more successful local recruitment remains unclear [14], in part due to lack of prescreening data.

Systematizing and centralizing prescreening data capture is uncommon [15], in part due regulations that restrict formal data collection prior to consent and inclusion of prescreening data in parent trial databases as well as limited resources available to support this effort. Yet, to understand the effectiveness of recruitment strategies and potential sample bias, it is critical for trialists to assess the full recruitment process. This process begins with efforts to increase awareness and interest in trials and data are needed to examine this “top of the funnel” (Fig. 1). Limiting recruitment data to those collected after consent at in-person screening visits tells only part of the story.

Fig. 1
figure 1

Clinical trial recruitment “funnel”

To evaluate recruitment prior to enrollment, we designed and developed a centralized prescreening database, the data-driven approach to recruitment (DART), for an ongoing preclinical AD trial conducted by the NIA-funded Alzheimer’s Clinical Trials Consortium (ACTC). In this manuscript, we describe the design of the DART database and the share pilot data obtained from the initial vanguard phase used to assess the feasibility of centralizing prescreening data collection.

Methods

The prescreening database was implemented under the AHEAD 3–45 study, a clinical trial evaluating the safety and efficacy of lecanemab (BAN2401, Eisai Inc.) in individuals who may be at increased biological risk for AD dementia [16, 17].

Design

The DART database was collaboratively developed by a working group of ACTC coordinating center personnel and participating study sites. This working group met monthly from September 2020 to June 2021, first to design the data collection form and then to establish methods to minimize barriers and maximize the likelihood of adoption by many study sites. We viewed the inclusion of site personnel in this working group as essential to the success of this initiative.

Variable selection

Variables selected for the DART database were aligned with the ACTC Minimal Data Set (MDS) recruitment and demographic variables (Table 1). To minimize site burden, we restricted the number of variables collected to eleven, including seven categorical and four free-text fields.

Table 1 Prescreening database variable description

Site selection and responsibilities

In the DART vanguard phase, seven active AHEAD 3–45 sites collected prescreening data for approximately eight months. Vanguard sites were selected with attention to balance across site type and experience, including experience in similar trials, existing infrastructure to capture prescreening data, and existing prescreening databases. Vanguard sites were reimbursed for their participation.

A key component of the vanguard phase was a monthly meeting with representatives from each site to discuss implementation and to share and review metrics generated from prescreening data. The goal was to use preliminary site experiences to identify opportunities to improve database design, reduce site burden, facilitate timely data entry, and improve data integrity.

Electronic data capture system (EDC)

We developed a separate EDC system specifically for this prescreening initiative, using the same framework as the AHEAD 3–45 study EDC [18]. Given that preexisting methods of capturing prescreen data varied widely across sites, we offered two options for sites to transmit prescreening data. For sites not capturing prescreening data electronically, data were entered directly into the EDC by site personnel. For sites with preexisting prescreening databases, batched upload was permitted, if it was performed at least every 2 weeks. A Data Transfer Agreement ensured the uploaded data were coded and formatted appropriately. Summary data reports were developed and distributed to sites and study leadership.

Institutional review board approval and informed consent

The central IRB governing the AHEAD 3–45 study (Advarra, Columbia, MD) determined that the prescreening database was of minimal risk and did not require a formal informed consent process. Advarra granted the study a Waiver of Consent and Waiver of HIPAA after determining that the waiver satisfied the Common Rule and the criteria set forth in the HIPAA Privacy Rule at (45 CFR 164.512(i)(2)). Only deidentified information is collected in the central database.

Statistical analysis

Time to contract execution and time to IRB approval for this initiative were defined as the time from when the site’s participation was confirmed to the time the contract was fully executed and central IRB approval was received by each site, respectively. Time to data entry was defined as the time the contract was fully executed to the time data entry/upload was initiated at the respective site.

Continuous variables were summarized by means and standard deviations while categorical variables were summarized using percentages. All statistical analyses were performed using R (Version 4.1.0).

Results

Each of the seven vanguard sites was able to provide data in this initiative. Six of the sites opted to directly enter their data into the EDC, and one site utilized the batched upload functionality. Mean time to IRB approval after site selection was 124.1 days (range: 62–157 days), and mean time to contract execution was 213.1 days (range: 167–299 days). Mean time to data entry was 47.6 days (range: 2–128 days) across sites.

Sites reported some challenges while initiating this protocol at their sites. The most common barrier reported was the inability to identify staff at sites to complete the data entry, primarily affecting time to data entry. Other issues raised included the handling of incomplete records (i.e., participants decided not to proceed with screening before demographic information was collected), difficulty tracking prescreening status and updating records accordingly, and inconsistent entry of the PTID# in the prescreen EDC after eligible participants attended a study screening visit. For the batched upload site, minor formatting issues arose that required correction prior to incorporating the data into the database. These included using incorrect separators (“/” instead of “|”) and incorrect coding of gender. Once these barriers were addressed, subsequent issues entering the data were minimal.

Table 2 displays the demographic summaries and recruitment source of participants prescreened during the vanguard phase by site. Most prescreened participants in this vanguard phase self-reported as being female sex, White race, and non-Hispanic ethnicity. The total number of prescreened participants varied widely among the vanguard sites (range: 3–611 participants), as did the recruitment sources of the participants. Referrals through websites (including study website, site website, and ClinicalTrials.gov) consistently produced a meaningful proportion 55.8% (range: 36.0–75.0%) of prescreen activity.

Table 2 Demographics and recruitment source by site

Table 3 summarizes recruitment source data by race and ethnicity, respectively. Websites were the most common recruitment source across racial groups. Hispanic participants appeared to have more often been recruited from registries and local recruitment efforts than to non-Hispanics participants.

Table 3 Recruitment source by race and ethnicity

Table 4 presents eligibility results, reported by race and ethnicity. During the vanguard phase, 19% of prescreened participants had been deemed eligible for in-person screening, though many prescreened participants remained in the prescreening process at the time of data freeze.

Table 4 Eligibility by race and ethnicity

Table 5 shows the distribution of reasons why participants did not continue to in-person screening by race. The most frequent reason was a loss of interest or concern about study burden, though many participants (50%) who did not proceed to in person screening had reason entered as “other.”

Table 5 Reason for prescreen fail by race

Qualitative feedback collected from the vanguard sites through the monthly meetings suggested that the metrics routinely shared were helpful in guiding local recruitment efforts. Therefore, site-specific reports will be generated for all participating sites, and study-level web reports will be made available in real-time to study leadership for the study-wide implementation phase of this initiative.

Discussion

Through the DART initiative, we demonstrated that the collection of prescreening data in a multi-site clinical trial is feasible. Seven vanguard sites were able to enter or upload prescreening data into an EDC developed specifically for this purpose. We also demonstrated that meaningful questions can be answered by capturing key variables from the prescreening phase.

From these preliminary data, we identified recruitment strategies that more often yielded prescreened participants than others. We also observed early trends that the effectiveness of recruitment sources may differ among racial and ethnic groups. For example, multiple sources of outreach including local campaigns, such as local television or radio interviews, account for a slightly higher percentage of Hispanic prescreens, compared to non-Hispanic, though the sample size remains low. Moreover, the AHEAD 3–45 study website accounted for a high percentage of prescreened participants across several racial and ethnic groups. Given that most of the central recruitment strategies implemented for the study promoted the study website, this suggests that these efforts may be a critical element toward improving demographic representation in this study.

Measuring prescreen failure rates may offer important guidance to trial leadership. From these vanguard sites, we found that losing interest and the unwillingness to endure trial burden were more frequent reasons for people to not proceed to in-person screening than was ineligibility based on trial enrollment criteria. This finding could help inform changes in recruitment materials or site practices to explore means to reduce burden or to make research participation more appealing to potential participants. Notably, had trial enrollment criteria been a primary reason for failure to advance to in-person screening, such data would provide the study team the opportunity to review and potentially revise the trial inclusion/exclusion criteria.

Next steps

As we move towards study-wide implementation phase of DART, some changes have been made to the data collection form. As noted above, a high percentage of reasons participants prescreen fail were entered as “other,” followed by a free-text description. In response to this, we expanded the options that sites can select for “Reason for Prescreen Fail.” We used the reasons written in the free-text field to expand the categories to match the exact inclusion/exclusion criteria from the trial, including “age,” “does not have additional risk factor (< 65 years old only),” “already enrolled in another clinical trial,” “no longer interested—lives too far from study site,” and “lost to follow-up/unable to contact.” The addition of these options will help limit the number of “other” reasons in the expanded initiative and permit more useful and analyzable data. We also decided to eliminate the collection of occupation and education as these variables were infrequently collected in the initial stages of the prescreening process and hence resulted in substantial missing data.

Using these methods to centralize prescreening data collection requires effort from site personnel, project management, data management, and biostatistics, making funding an essential component for success. Ideally, resources to collect prescreening data would be included in the original study budget. Some but not all trials offer start-up funds for the effort of securing IRB approval and other preparatory needs, as well as recruiting participants for initial screens. Inclusion of the site effort to put a prescreening database and infrastructure in place might ideally be included as a line-item in start-up budgets. Alternatively, the resources for maintaining prescreening databases might be provided as part of infrastructure resources for new or established trial site consortia.

Limitations

We acknowledge some important limitations. The DART initiative vanguard phase utilized only 7 sites from the AHEAD 3–45 study with a small number of data variables. This was done to limit site, participant, and coordinating burden and to enable collection of preliminary experiences with the initiative. The prescreening initiative may result in duplication of effort for at least some sites that already capture prescreening data electronically, though we offered the batched upload option to minimize burden for those sites. The initiative has costs, which may limit the ability for small and/or underfunded trials and trial networks to create this infrastructure, potentially limiting the generalizability of this effort. Alternatively, collecting these data may enable efficient use of recruitment resources, potentially reducing overall trial costs. The initiative started early in the recruitment phase of the AHEAD 3–45 study, which may have had an impact on the number of prescreens sites entered. This may accurately reflect start-up in future trials, but we have limited information related to the main stages of study accrual. The COVID-19 pandemic may also have influenced these results, given the impact on site staffing during the vanguard phase and possible effects on willingness to participate in the AHEAD 3–45 study. Finally, though the study website yielded the most prescreens, it is unclear how participants found the study website as other advertisements may have directed them towards the website. Though this is a limitation, it does support the utility of a study website as a mechanism for potential participants to connect with sites.

Conclusions

Recruitment for clinical trials is challenging and time consuming. Relying on post-consent screening data is insufficient to fully capture the effectiveness of centralized and local efforts to accrue a full sample and identify sources of selection bias. The centralized collection of prescreening data may increase the efficiency, speed, and effectiveness of study recruitment, including enrolling a cohort more representative of the population at large. The vanguard phase of this innovative prescreening database initiative demonstrated the feasibility of establishing such a database and allowed the project team to learn important lessons to increase the likelihood of a successful study-wide implementation.