Background

The recruitment and retention of participants to health research presents challenges for researchers. A randomised controlled trial (RCT) that is embedded within a host research study, testing one or more alternative research designs, is a suitable way to examine which aspects of research design are most effective: such an approach has been suggested for improving response to research questionnaires [1, 2] and enhancing trial recruitment [3,4,5]. Embedded trials can also be useful in identifying the most effective methods of retention [4]. We define retention as the process of keeping a participant involved in a research study after they have agreed to take part.

Postal questionnaires are widely used as a method of data collection in health research and are a cost-effective option compared with telephone and email [6]. Non-response to questionnaires reduces the available sample size, decreases the precision of parameter estimates and can introduce non-response bias [1, 7]. There is a long history of research on methods to improve survey response [8, 9]. There is strong evidence that financial rewards (and to a lesser extent financial incentives) can improve questionnaire response rates [10], even when the amounts are small [11], and that shorter questionnaires [12,13,14] and reminders [13,14,15] are often effective. Response is also conditional on study-related factors including the organisation sending the letter and how the recipient regards them [8, 9]. Of particular relevance to this study, personalisation of the cover letter can improve response, such as adding a wet-ink signature [16, 17] or addressing the respondent by name [16, 18]. Personalisation is not always effective [5, 18]. This variation in effect is influenced by the competing effects of personalisation, which seeks to build a connection with the recipient, and anonymity, which tries to achieve the opposite, creating a ‘two-edged sword’, particularly in surveys requesting sensitive information [8].

One variant of personalisation is social pressure, which involves persuading people to behave in a certain way by the promise that their actions will be made public [19]. It arises because we are all eager to be seen in a positive light by those around us. Public praise and shame make a behaviour more salient and can encourage citizens to comply with a request [20]. Pride arising from acclaim by peers can motivate people to persevere in carrying out actions, despite obstacles [21]. In economics, ‘image motivation’ describes how citizens seeking social approval may choose to exhibit qualities that they think are widely regarded as good. Research indicates that people are more likely to act in a prosocial way in a public space than in a private space [22] and that charitable donors enjoy the feeling of ‘prestige’ they experience when their donations are publicised [23]. In social sciences, ‘social pressure’ can be effective in raising voter turnout in elections [19, 24] and in encouraging charitable donations [25, 26]. Social pressure conveys to the participant that information about their behaviour will be noticed. It can be done in a way that either engenders pride or shame, although in health research the utilisation of shame is likely neither desirable nor ethical. Social pressure can also be regarded as a social incentive, which informs the recipient that a verbal or non-verbal reward will be delivered if there is progress in performing the behaviour [27].

Social pressure is unlikely to be effective among those who are very committed to the behaviour or those who are not at all interested: it is particularly suited to those who have undertaken the behaviour before or are already thinking about the behaviour, but in need of persuasion or reminder [24]. This makes it particularly suitable for a longitudinal health survey, where the participants have already completed an initial questionnaire and indicated willingness for future participation.

We undertook an RCT to test the application of social pressure to questionnaire retention. We implemented a mild form of social pressure, letting participants know that their behaviour was being monitored [28], which we thought would be more appropriate in this context than stronger versions of social pressure, which promise to advertise the behaviour in a public place. The objective of the embedded trial was to test the effect of notifying survey participants that their responses are being noticed compared to a simple letter control, on response to a longitudinal health survey. We follow the CONSORT trial reporting guidelines [29,30,31], including the extension for embedded recruitment trials [4]. The protocol is available from the lead author.

Methods

Trial design

We implemented a parallel group RCT which was embedded in a longitudinal health survey. Participants were randomly allocated 1:1 to receive standard covering letters (control condition) or the same covering letter with additional wording to convey that participant responses were being noticed. The letters were sent to accompany two consecutive cohort questionnaires, at six and 12 months after baseline; each letter was posted simultaneously to all participants in both experimental groups. The primary outcome was response to both surveys, to test for the effect on retention. The trial was conducted between July 2015 and April 2016 in Salford, UK.

Host study

The Salford Integrated Care Programme (SICP) is a large-scale integrated-care programme to improve care for older people with long-term conditions and social-care needs in Salford, a city in northwest England which has high levels of deprivation and long-term illness. The Comprehensive Longitudinal Assessment of Salford Integrated Care (CLASSIC) (the host study) is a cohort study, designed to provide a rigorous test of the ability of the SICP to deliver improved care for older people. The CLASSIC cohort was recruited and followed over time, to assess the overall impact of SICP, while subgroups of the cohort were used to evaluate different SICP interventions.

All general practices were approached to participate in the CLASSIC study and 33 out of a possible 47 were recruited. Inclusion criteria were patients aged over 65 years who were on practice registers for at least one of the following long-term conditions: atrial fibrillation; high blood pressure; coronary heart disease (CHD); heart failure; stroke, chronic obstructive pulmonary disease (COPD); asthma; chronic kidney disease (CKD); diabetes; epilepsy; psoriasis; rheumatoid arthritis’ and osteoarthritis. CLASSIC employed an innovative recruitment approach using FARSITE, which enables researchers to search anonymised GP health records. FARSITE was used to manage and create lists of eligible patients and then each practice was asked to identify patients meeting the exclusion criteria (patients in palliative care and those with conditions which reduce capacity to consent and participate). Once approved by GPs, a link with a third-party service enabled the invitation letters to be mailed remotely. Practices did not receive incentives to take part but did receive support costs to reimburse their time.

Participants in the host study

The target population for the host study were people aged over 65 years with a long-term condition. The host study recruited 4447 people, who consented to receive four questionnaires on a six-monthly basis. This was out of 12,989 eligible patients who were sent an initial baseline questionnaire in January 2015. The eligibility conditions for the embedded study were the same as the host study and all host study participants were included. The flow of participants through the host and embedded studies is illustrated in Fig. 1.

Fig. 1
figure 1

Flow chart of participants in the host study and embedded study

Participants were sent four postal surveys, at baseline and then every six months thereafter. The surveys in which this RCT was embedded were the second and third surveys, which were sent at six and 12 months after baseline. The surveys were similar to one another and contained brief measures of service experience, health and care outcomes and service utilisation. These measures were linked to routine data from electronic records on clinical parameters, medication use and interactions with NHS and social care services. Maximising retention across all four surveys was crucial for the validity of the host study. The second questionnaire was sent in July 2015 to those who had completed the first questionnaire. The third questionnaire was sent in January 2016 to everyone in the cohort, except those who actively withdrew or died. At each stage, participants who did not return a questionnaire were sent a second copy with a reminder letter three weeks later. Participants in both groups were offered an incentive of a £10 voucher for completion of the first (baseline) questionnaire and £5 for completion of the third questionnaire, but no incentive was given for completion of the second questionnaire.

Intervention

All participants were sent a covering letter and a copy of the CLASSIC questionnaire. The content of the intervention and control letters was similar (Additional file 1) and the intervention letter with the first questionnaire included the following additional text. The wording of the intervention was designed to exert social pressure by letting the recipient know that their previous response had been noted and that future responses would be noted by the researcher and communicated to the recipient at the time of the subsequent questionnaire:

figure a

The intervention letter with the second questionnaire included the following additional text, which was designed to exert further social pressure by letting the recipient know that their previous two responses had been noted and that future responses would be communicated to the recipient at the time of the subsequent questionnaire:

figure b

Participants in the control group received the standard covering letter, similar to the intervention group, but without the additional text. Both intervention and control groups received the same encouragement from the researchers on the host study to return completed questionnaires, including financial incentives and follow-up phone calls.

Outcomes

The objective of the intervention was to improve retention in a cohort study, so we focus the primary outcome on completeness of response. We were interested in ‘unit’ response, defined as whether the respondent returns the survey, rather than ‘item’ response, defined as whether the respondent completes particular questions on the survey. The primary outcome for the embedded trial was retention in the cohort, measured by the return of both completed questionnaires. The definition of a returned questionnaire was that adopted by the host study: any questionnaire with at least one completed scale. Secondary outcomes were: return of a completed first questionnaire; return of a completed second questionnaire; and return of either questionnaire.

The original protocol was concerned with the first questionnaire and specified the primary outcome as unit response to the first questionnaire. The primary outcome was changed to ‘unit response to both questionnaires’ on the recommendation of the chief investigator of the host cohort study prior to him having access to the response data to the second questionnaire. This was viewed by him as the most important endpoint with regards to retention in a cohort study.

Sample size

A conventional sample size calculation was not appropriate, as the sample size was determined by the size of the host study (the 4447 individuals who had already been recruited to the CLASSIC cohort and returned a baseline questionnaire). We nevertheless present some estimates of power given likely effects. The cohort included 380 pairs of individuals living in the same household, most of whom were couples. We expected that individuals in the same household were likely to have sight of each other’s letters and we wanted to avoid contamination between different experimental conditions, so we ensured that all members of a household were assigned to the same intervention. The available sample of 4447 individuals provides 80% power to detect a difference of around 4.5 percentage points between a control group response rate of 70% and a treatment group response rate of 74.5%, which equates to an odds ratio (OR) of around 1.2 (two-sided, p > 0.05), which is similar to the effect of a similar letter on voter turnout [28]. We have not accounted for clustering by household in the sample size calculation, because the effect was anticipated to be small, but we have taken account of it in the analysis.

Randomisation

Randomisation was undertaken at the household level and stratified by household size to ensure that the multiple occupancy households were evenly distributed between the two groups and to keep the groups approximately equal in size. A random number variable was generated, uniformly distributed between 0 and 1. The single-person households were randomly assigned to treatment and control groups, using blocking to achieve groups of equal size. This was repeated for the two- and three-person households. Randomisation was undertaken using Stata 14 [32].

Blinding

Participants were blinded to their participation in the embedded study. They were not informed that other households were being sent differently worded letters: we expected that their actions might change if they knew what others receive. This approach was acceptable to the research ethics committee. Randomisation was performed after recruitment to the host study, ensuring allocation concealment. The research team were not blinded to the intervention, but had minimal contact with trial participants. Researchers were blinded to group assignment during outcome ascertainment.

Statistical analysis

Data management followed the existing procedures for the host study. Possible dispositions [33] were: returned questionnaire; died; withdrew (participants were advised that the return of a blank questionnaire would constitute withdrawal); not returned; and non-contact. For the primary intention-to-treat (ITT) analysis, the disposition is categorised as returned or not returned (which includes died, withdrew, not returned and non-contact). We present summary descriptive statistics of participants by group. We estimate the effect of group allocation on unit response, using logistic regression. To control for multiple participants per household, bootstrapping with adjustment for clustering was used, with 1000 replications. To adjust for any baseline imbalance, covariates for age (over 75 years), poor quality of life and male gender, living alone and having no qualifications were included in the model. All analyses were carried out in Stata 14 [32]. The main analyses were performed on an ITT basis, including all randomised participants, whether or not they received a questionnaire. We conducted a sensitivity analysis where participants with a missing address were excluded.

Results

The first questionnaire was posted in July to August 2015 and the second in January to February 2016. The flow of participants through the trial is shown in Fig. 2. All 4447 participants who had responded to an earlier questionnaire were randomised to receive the social pressure letter (n = 2223) or a control letter (n = 2224). Questionnaires were not sent to people without an address on record and those who were known to have died or withdrawn from the host study. No participants were lost to follow-up in the embedded retention trial: all 4447 participants provided an outcome measure simply by returning or not returning the questionnaire (Fig. 2).

Fig. 2
figure 2

Consort flow diagram

The baseline demographic characteristics and health status of trial participants are presented in Table 1. The mean age of the trial population was 75 years, just over half were women (54%) and they were largely of white British ethnicity (95%). Half of the sample had no qualifications and one-fifth needed help with reading materials. All the participants had long-term health conditions; on average, each person reported five health conditions [34]. Quality of life was rated as good or very good by 71% of participants; 52% said they were satisfied with their health, measured by WHOQOL [35]. Characteristics at enrolment were balanced between the two groups.

Table 1 Participant characteristics and health status at baseline, by group

All 4447 randomised participants were included in the primary ITT analysis. The outcome was a binary measure of whether a questionnaire was returned. We are not aware of any participants receiving a different intervention than the one they were allocated.

A summary of all observed outcomes and treatment effect estimates is given in Table 2. The primary outcome is return of questionnaires 2 and 3. Both questionnaires were returned by 1577 participants (71%) who were assigned to the social pressure letter and 1511 (68%) assigned to the control letter, an increase of 3 percentage points. The OR (95% confidence interval [CI]) is 1.15 (1.01–1.31). This effect is sustained with an or (95% CI) of 1.16 (1.02–1.33) after adjusting for clustering within households and factors thought to affect response: age; quality of life; gender; living alone; and education. All the secondary outcomes, questionnaire return at six months, questionnaire return at 12 months and return of at least one questionnaire, showed no statistically significant difference between the two groups, although for all outcomes, the response was higher in the social pressure group than in the control group.

Table 2 Results: questionnaire return by experimental condition, using logistic regression adjusting for household and baseline characteristics and clustering

In a sensitivity analysis, we repeated the analysis excluding 152 randomised participants who were not sent either questionnaire because their address was unavailable. Excluding these participants from the unadjusted analysis of the primary outcome leads to an OR (95% CI) of 1.14 (1.00–1.30) so the conclusions are robust regardless of whether we exclude these participants. It is usual practice to report the intracluster correlation coefficient (ICC) in cluster trials, but in this trial there is so little clustering that the ICC would be uninformative: 3642 clusters (90.1%) were single-person households and only 402 clusters (9.9%) were households of two or more persons.

Discussion

This evidence suggests that the addition of a simple social pressure intervention to a covering letter may lead to a modest increase in the number of participants returning more than one questionnaire. Use of social pressure interventions of this kind can aid retention in health cohort studies and the effect is similar to that found for other methods of personalising covering letters [18]. The response to the intervention was similar at both time points, with no sign of a diminished effect. It would be feasible to deliver social pressure to a cohort over a longer period, but we have no evidence on whether it would continue to be effective. It is plausible that the effect might wane, particularly with repeated non-responders receiving letters confirming their history of non-response. The only additional cost to run the study was £380 for the postal company to assign the correct letter to each respondent. The intervention requires minimal cost and effort and even modest effects of the size seen here may justify their implementation. Modest interventions, such as social pressure, are never likely to have large effects, but a combination of best practice modifications like this may combine to show more substantive benefit [36]. Fifteen years on, we can echo Don Dilman’s call for more research on how different approaches can complement one another and how they vary by population and topic [9].

The study population were older adults, with multiple long-term conditions. This older cohort, in common with people of this age group across the UK [37], had a relatively low level of formal education. Further investigation of the intervention in different situations such as longer-term cohort studies, cohorts with different patient groups, other age groups and in randomised trials is warranted.

The intervention implemented here was a very mild form of social pressure, based on highlighting to the recipient that their actions are being noticed by the sender of the questionnaire. The modest success of this mild form may suggest that investigation of interventions using stronger social pressure messages is warranted. Stronger social pressure in this context could include inducing pride by offering in advance to publicise the details of those who respond to the questionnaire, for example by naming consenting responders in a study newsletter. There is likely to be an optimal amount of pressure in this context: very strong social pressure used in other settings, such as shaming people for non-response, would be counter-productive and unacceptable in a health context. The acceptability of social pressure interventions will likely vary according to population type, questionnaire topic and country regulations, for example, social pressure may fall foul of the ‘common rule’ (45 CFR 46) in the United States [38].

The benefits of undertaking an embedded trial of this type are that the financial cost and time effort were minimal over and above those of the host study. Some additional work was involved in allocating people to different interventions and in sending out alternate versions of the letter, tailored to group allocation and nature of their response. No additional recruitment, consent or data collection were required, making effective use of participants’ time. This embedded trial methodology provides a valuable opportunity for rapid investigation of novel retention interventions, 12 months from first idea to results in this example. There is minimal ascertainment bias as we got ethical approval that participants were unaware of the study.

The study design presents some minor challenges, the main one being the constraints on sample size, which, unlike embedded recruitment trials, cannot exceed the sample size of the host study. Researchers on the embedded trial are reliant on the host study for data collection and management, and may have little control over the timing and methods. The design of the host study may introduce analysis issues, such as in this example, the clustering of participants in households. These all proved to be minor limitations, compared with the benefits.

Conclusions

The addition of a simple social pressure intervention to a covering letter accompanying a health cohort questionnaire leads to a modest increase in the number of participants returning more than one questionnaire. The intervention requires minimal cost and effort, and consequently even modest effects of the size seen here may justify their implementation. An embedded RCT is a practical and efficient method of conducting research on the effect of behaviour change interventions on retention.