Background

Some recent systematic reviews suggest that trial participants who are allocated to placebo groups experience adverse events (AE), including AEs attributed to apparent drug interactions [1,2,3,4,5,6]. Yet adverse drug reactions cannot be directly caused by placebo treatments. There are two overlapping explanations for how this might occur. First, a patient may have an underlying condition whose natural history produces some event (such as a headache), then the patient misattributes the event to the trial intervention (in their case, a placebo). Second, having been warned about side effects in the patient information sheets (or elsewhere), the patient may expect an AE. This negative expectation could then produce the event [7]. There is some empirical support for the latter explanation. In one multicenter randomized trial of aspirin or sulfinpyrazone for treating unstable angina, due to differences in individual hospital review processes, patients either received or did not receive a statement outlining possible gastrointestinal side effects. This resulted in a sixfold increase (P < 0.001) in the number of individuals withdrawing from the study because of subjective, minor gastrointestinal symptoms [8]. Major (“objective”) complications such as peptic ulcer or bleeding as diagnosed by study physicians were similar across centers.

There are two gaps in the literature about nocebo effects within clinical trials. First, the systematic reviews in the area have not been synthesized. This leaves questions about the comparative prevalence of AEs within placebo groups across different conditions unanswered. Second, the existing systematic reviews often conflate natural history with nocebo effects. The fact that an AE occurs after taking a placebo does not imply that the AE was caused by the placebo. This mistaken inference was noted in studies investigating positive placebo effects > 20 years ago [9]. It was resolved by comparing what happens within three-armed trials (treatment, placebo, no treatment) to patients who receive a placebo treatment with patients who are left untreated. The same method could also be used to differentiate between nocebo effects and natural history.

Objectives

In this rapid systematic review, we addressed both of these gaps and:

  1. (1)

    synthesized the systematic reviews of nocebo effects; and

  2. (2)

    reviewed trials within these systematic reviews that reported AEs in untreated control groups to compare these with AEs in placebo groups.

Methods

Protocol and registration

The protocol for this rapid review was published on PROSPERO on 05/04/18 (record no. CRD42018092437).

Eligibility criteria

We included any systematic review of randomized trials where AEs were reported and quantified. We included systematic reviews of trials in any condition. We excluded reviews of non-randomized trials, and reviews that did not include quantitative data about AEs within trial control groups.

Information sources

We searched for PubMed records with the word “nocebo” in the title and “systematic” in any field (see Appendix). We also contacted experts and searched references of included studies.

Study selection and data collection

Two authors (JH, NK) extracted data from the systematic reviews: about year of publication, authors, disease area, type of AE, and rates of AEs.

Two authors (NK, RW) extracted data from the trials within the systematic reviews to identify those which included AE rates within untreated groups and compared these with AEs (in the same trial) in placebo groups. They reported AE rates in all three groups where these were found. One author (JH) acted as a second extractor for a random selection (n = 50) of the studies. Discrepancies were resolved by discussion between review authors.

Risk of bias in individual studies

To ensure a minimum level of quality within included reviews, we assessed whether the following two items of the Assessment of Multiple Systematic Reviews tool (AMSTAR) were satisfied: were two or more electronic sources searched?; and was the scientific quality of the included studies assessed and reported? Other overview authors have used similar assessment criteria [10, 11]. We performed a sensitivity analysis excluding studies which did not meet the two criteria.

Summary measures

We summarized the median and interquartile ranges (IQR) of AEs, as reported by review authors, as well as drop-out rates. When at least three studies reported specific AEs (such as headache), we reported these, also using medians and IQRs.

In trials within the reviews that reported AE rates in untreated groups, we summarized the rates of AEs in the untreated and placebo groups. To complete this analysis, we needed references to individual trials within the reviews. When such references were not published in the systematic review, we sent the corresponding author an email request. If we did not receive a response, we sent a follow-up email, after which (if we still did not receive a response) we did not search the trials within that review.

Synthesis of results

We reported the median and IQRs for overall AEs and dropout rates due to AEs. For our analysis of AEs reported in three or more studies, we also reported medians and IQRs. Among the trials within the review that included untreated groups, we compared the prevalence of AEs in untreated groups with the prevalence of AEs in placebo groups.

Other potential sources of bias

A source of potential bias could have been the method by which the AEs were ascertained. AE evaluation strategies involving structured assessments (checklists of likely AEs) risk confusing misattribution of symptoms caused by the underlying condition to the (placebo) intervention. For example, pain, nausea, and headaches are common symptoms of many diseases. Patients who are asked whether they experienced one of these are therefore likely to say yes because of the underlying condition or simply because these are common. We could not infer from the report whether such an event (when ascertained using structured assessments) was caused by the underlying condition, a negative expectation, or a treatment. Spontaneous strategies for measuring AEs address this problem by not naming potential AEs in checklists and instead asking patients to report any AEs that arise. We compared AEs assessed using the different methods in an exploratory analysis. Data were not suitable for pooling for this analysis so we reported our findings narratively.

Results

Study selection

Our search yielded 20 systematic reviews (see Fig. 1) that met our inclusion criteria [12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31]. A total of 1271 randomized trials and 250,726 placebo-treated patients were included within these reviews.

Fig. 1
figure 1

PRISMA diagram records

Study characteristics

The studies were all published in the last 12 years and pain-related conditions were most commonly studied within the reviews [12,13,14, 16, 19, 20, 23]. Other conditions were: depression [18, 21, 24, 26]; multiple sclerosis [22]; motor neuron disease [27]; restless leg syndrome [28]; Parkinson’s disease [29]; Alzheimer’s [30]; cardiovascular disease [25]; and epilepsy [31]. Two reviews included trials in any condition [15, 17].

Some of the reviews reported data separately for different types of treatments or conditions without combining the data. Amanzio et al. [12] reported trials of non-steroidal, anti-inflammatory drugs (NSAIDs); triptans, and anticonvulsants separately; Häuser et al. [13] reported data for fibromyalgia and painful diabetic neuropathy trials separately; Mitsikostas et al. [19] reported data for symptomatic and prophylactic trials separately. Papadopoulos et al. [22] reported symptomatic and disease-modifying trials separately, and Rief et al. [25] reported tricyclic (TCA) and selective serotonin reuptake inhibitor (SSRI) trials separately. We followed the authors and did not combine the separately reported data.

Ninety distinct types of AEs were reported. These ranged from abdominal pain and dry mouth to cancer and stroke.

Risk of bias within studies

Eight reviews [12, 13, 15, 16, 18, 25, 26, 28] met both of our quality criteria (searched at least two databases and reported data about quality of included trials, see Additional file 1: Table S1). Seven reviews met neither quality criteria [19, 20, 22,23,24, 27, 30]. Of the five studies that met one of our quality criteria, one [14] searched more than one database and four assessed and reported the quality of included studies [17, 21, 29, 31]. There were insufficient data to formally analyze the relationship between review quality and AE data; however, there did not seem to be any correlation between review quality and AEs. Studies which met at least one criteria had the highest rates of AEs (64.7%) compared with studies that met both criteria (36.6)%) and no criteria (54.9%), while the rates of drop out were similar across those which met no (6%), one (4.8%), and both (5.3%) quality criteria.

Results of individual studies and synthesis of results

The results of individual studies are provided in Table 1. The median prevalence of AEs in trial placebo groups (as defined by the individual trials within the reviews) was 49.1% (IQR 25.7–64.4%). The median rate of dropouts due to AEs was 5% (IQR 2.28–8.4%).

Table 1 Summary of adverse effects in placebo groups within trials

Only some reviews provided details about specific AEs reported within the reviewed trials. In all, 90 different AEs were reported and 17 types of AEs were reported in at least three trials (see Table 2). Eight of these had median rates > 5%: headache (18%); vomiting (7.7%); fatigue (7.11%); insomnia (5.7%); burning (5.88%); somnolescence (5.56%); dizziness (5.1%); and constipation (6.43%).

Table 2 Adverse event (AE; %) within placebo groups, by type of AE (where at least three reviews reported the same AE)
Table 3 Adverse events (AE) in placebo groups compared with AEs in untreated groups

Fifteen trials within the reviews contained data about AEs reported in untreated groups (see Table 3). The average AE rate in these trials was smaller than the AE rate in placebo groups and treatment groups within the same trials (4.25%, 6.51%, and 5.93%, respectively). The AE rate in the placebo groups was higher than that of the untreated groups, although the difference was not statistically significant (P = 0.07), tentatively suggesting that the AE rates cannot be attributed to natural history. One study reported AEs within untreated groups, but not in a format we could pool [32]. In it, the authors found no significant differences between AEs reported in untreated, placebo, or treatment groups.

Other sources of bias

Structured assessment seemed to result in different AE reports in some of the reviews (see “Additional analyses”).

Additional analyses

Two reviews checked for differences between included trials with high versus low quality [12, 29]. Amanzio et al. [12] found that of 28 types of AEs measured, only one (somnolescence) showed a difference. Stathis et al. [29] found that the higher the trial quality (measured by Jadad scores), the lower the dropout rate in placebo groups.

Four reviews reported data about structured versus spontaneous reports of AEs [12, 14, 25, 26]. Only one of these found a statistically significant difference between AEs reported by different methods: Rief et al. [25] found it to be more than twice as likely for an AE to be reported using structured as opposed to spontaneous assessment methods (odds ratio 2.6, 95% confidence interval 2.1–3.3). For the 24 types of AEs where it was possible to compare spontaneous versus structured assessment, Amanzio et al. [12] did not find statistically different rates of AE reporting in 21 of them. Among the AEs where there was a difference, structured assessments resulted in higher prevalence of AEs for nausea in NSAID trials; spontaneous assessment resulted in greater AEs for fatigue in NSAID trials and nausea in Triptans trials. Hauser et al. [14] planned to compare AE rates for different assessment strategies but decided post hoc not to, due to poor quality of reports of assessment strategies; there did not appear to be a difference. Rojas-Mirquez et al. [26] reported which trials used different assessment methods but did not say whether there was a difference (and none were obvious by visual inspection).

Discussion

Summary

Our overview of systematic reviews of nocebo effects suggests that almost half of patients in the placebo groups within clinical trials experience AEs that are attributed to the drug. One in 20 patients in placebo groups drop out due to drug intolerance. This is higher than similar reports of AEs reported by patients taking the same interventions in routine clinical practice [14, 25].

Comparison with other evidence

Our study adds to the existing reviews showing that AEs are common among patients in placebo groups within clinical trials. We showed that this phenomenon is common across several conditions and extends to many types of AEs. We also added a comparison of AEs in untreated groups, which suggested that natural history or misattribution is unlikely to explain the AEs reported within the placebo groups.

Strengths and limitations

This is the first synthesis of AEs that covers both different disease areas and different types of placebos. It is also the first to investigate natural history as a potential cause of AEs within trial placebo groups. Our conclusions are limited by the quality of the included systematic reviews and the quality of the trials within the systematic reviews. This limitation is mitigated by our quality assessment of the included systematic reviews, which suggests that they meet a minimum quality standard. Another limitation is that a second extractor only reviewed a random selection of studies independently. This is unlikely to have had an important influence on the results as there were only five discrepancies, mostly about the nature of follow-up in untreated groups within individual trials.

It was also difficult to disentangle different possible causes of AEs within placebo groups. While our analysis did not reveal any clear trend in AE reports by ascertainment method (structured versus spontaneous), it remains possible that the way patients are asked to report AEs could influence what they report. This compounds the problem of distinguishing between effects of negative expectations and the effects of mistakenly attributing routine symptoms to a trial intervention. Hence, future experimental studies where expectations are manipulated are likely to be required to determine the potency of negative expectations. Relatedly, the included reviews did not define AEs, perhaps due to the failure of the included trials to do so.

Implications for clinical trials and clinical practice

There are three implications of this study for trial methodology. First, methods to reduce the risk of AEs induced by negative expectations (that may arise due to the way possible harms are communicated) warrant investigation [33]. Some studies have already begun to address how clinicians might reduce nocebo effects within clinical trials, such as limiting suggestion of symptoms [34]. Identifying a model of informed consent that respects patient autonomy yet does not introduce unnecessary harm is therefore required. This might be achieved by shortening the informed consent process so that rare harms are made clear (for example, by being listed on a web page for patients who would like to know) but not forced upon patients. Oxman et al. proposed such a model [35], but it has not yet been implemented. Other ethicists have also called for “contextualizing” informed consent (adapting it for individual patients) [36]. Contextualized consent may enable practitioners to present benefits and harms to patients in ways that do not induce unnecessary nocebo effects. Second, we should modify the way AEs are collected such that natural history is not confused with AEs. This might be achievable with a combination of spontaneous and structured AE assessments.

Third, the inferences about intervention AEs needs to be rethought. This is typically calculated by comparing AE rates in treatment groups with AE rates in placebo groups. But if the AE rates in placebo groups are inflated by the way potential AEs are communicated to patients, this can lead to an overestimation of AEs in placebo groups and a consequent underestimation of AEs attributed to experimental treatments. The way AEs in trials are interpreted should therefore reflect the findings from this review.

Finally, our review also raises the ethical issue of balancing the need to respect patient autonomy and provide full informed consent, while at the same time reducing unnecessary harm done to patients. The former seems to demand that more information about AEs is provided, while the latter suggests that such information should be carefully communicated to avoid harms.

Conclusions

Trial participants who receive placebos within clinical trials report experiencing AEs. Research on clinical trial methodology is now warranted to reduce this potentially unnecessary harm.