Introduction

Oral immunotherapy (OIT) was first trialled as a potential disease-modifying treatment for IgE-mediated food allergy in 1908 [1], but until recently, there had been a relative hiatus in establishing OIT as a viable treatment modality. There is proven efficacy from numerous phase 2/3 trials where outcome measures have largely centred on the change in reaction threshold at supervised food challenges using incremental doses and (in some studies) improvements in health-related quality of life measures, although, typically, these are parent-reported. Although the key mechanisms of OIT are poorly understood, OIT has become increasingly offered within clinical practice in several countries [2, 3], though details remain limited in many countries. [4]. More recently, larger phase 3 studies are also now underway, as are studies looking at potential alternative routes of administration [5,6,7]. There is increasing interest from the commercial sector to develop immunotherapy products for food allergy, given public demand.

Allergen immunotherapy carries an inherent risk of allergic reactions, as dosing regimens need to balance a level of allergen exposure sufficient to induce immune desensitisation but low enough not to trigger an IgE-mediated reaction. The most important knowledge gap—and arguably the main barrier to regulatory approval—remains the safety profile (both short and longer term) of OIT in the treatment of IgE-mediated food allergies [8]. Studies report a rate of allergic reaction affecting between 18 and 100% of patients on active OIT, with 0.8–43% of doses causing symptoms [9,10,11]. This may lead to OIT failure, with 5–36% of patients being withdrawn during treatment [11, 12]. Immunotherapy has the potential to cause fatality through a severe adverse reaction, either during the initial screening challenges prior to treatment or dosing-related reactions [13].

With such wide variations in safety outcomes, being able to predict those at risk of greater and more severe adverse events is important. These factors can be divided into:

  1. 1)

    Those that are patient-dependent, where selective characteristics place a specific individual at higher or lower risk of AEs.

  2. 2)

    Features that are related to the protocol, that may affect the likelihood of dose-related AEs across study participants.

OIT efficacy has been evaluated in a number of systematic reviews, but analyses of dose-related adverse events (AEs) and other safety outcomes—both in relation to patient-dependent factors and those inherent to the protocol—have proved more problematic due to lack of consistency in reporting adverse events [14]. A meta-analysis by Nurmatov included 21 clinical trials of which 17 used OIT (with a predominance of studies assessing cow’s milk OIT), and found (perhaps unsurprisingly) that active OIT results in a greater risk of local and systemic allergic reactions compared to control treatments [15]. The authors attempted to evaluate the impact of different protocols and food allergens, but this was limited by the heterogeneity of reporting in the different studies. In consequence, only seven trials could be included in the analysis; the authors were unable to identify any factors associated with adverse events during immunotherapy, despite anecdotal reports suggesting milk and wheat immunotherapy often result in more frequent and more severe adverse events [16, 17].

In this article, we review the published data relating to the occurrence of dose-related adverse events and the inherent issues with such an analysis. Importantly, this needs to be framed in the wider context of our inability to predict reaction severity in food-allergic individuals [18] in a clinically meaningful way, and how this may be further confounded by the immune changes induced by OIT, resulting in unexpected breakthrough reactions [19].

Difficulties in interpreting safety data

Evaluating the differences in datasets generated by OIT studies in order to assess which factors might impact on safety is hampered by:

  1. 1.

    Heterogeneity in patients recruited: Differences in the eligibility criteria for study participants can be a major confounder in the analysis of safety data from studies. Some studies use a relatively high cutoff for allergen-specific IgE sensitisation as an inclusion criteria (e.g., 7 kUA/l [20]) compared with others which allow any level of sensitisation, something which might impact on the occurrence of AEs as there is some data to show that these are more common with higher levels of sensitisation. Historically, exclusion criteria have sometimes excluded those with a history of anaphylaxis [21], anaphylaxis (with respiratory symptoms) at baseline challenge [22]; most studies (probably quite reasonably) continue to exclude those with prior severe reaction requiring intensive care.

  1. 2.

    Heterogeneity in protocol design:

  1. (a)

    OIT protocols tend to consist of an initial updosing phase followed by a maintenance phase. More rapid updosing (e.g., “rush” updosing) or protocols where duration of updosing is shorter may be associated with a higher rate of AEs [8], although no head-to-head comparison has as yet been published.

  2. (b)

    Protocols vary in terms of the allergen “matrix” used to induce desensitisation: for example, peanut OIT protocols typically use defatted roasted peanut flour, but some studies have utilised roasted (not defatted) peanut in either ground or solid form, while others have used heat-modified peanut [23]. The different preparations may impact upon the occurrence of AEs, but thus confound any comparison between studies.

  3. (c)

    Many studies do not include a control group, and those that do may utilise routine care (i.e., allergen avoidance) rather than a true placebo control; studies incorporating a blinded placebo intervention tend to report higher rates of AEs than mere avoidance. While efficacy of active treatments should be compared with a placebo intervention, whether this holds true with respect to safety outcomes is less clear. For example, if a study reports a 50% rate of gastrointestinal (GI) events with active treatment, and 15% with a placebo intervention is the true rate of GI events nearer to 35% rather than 50%?

  4. (d)

    Studies also vary in the safety precautions taken, which impacts upon outcomes including occurrence of AEs. Some protocols include provision to temporarily cease dosing for fever or gastro-intestinal symptoms [22], while others mandate dose-reductions following dose-related symptoms. Thus, the occurrence of AEs will vary depending on the strategies used to manage dose-related symptoms during OIT, and may be clinician-dependent where the protocols provide flexibility to study investigators in terms of the available options.

  1. 3.

    Heterogeneity in AE reporting: Unfortunately, there is no current consensus on AE reporting across studies. Many studies only provide data relating to serious AEs requiring treatment with adrenaline [24], although adrenaline use is a poor marker of reaction severity [25].

    In contrast to studies of aeroallergen immunotherapy [26], there is no current consensus in terms of reporting structures for studies of OIT. Differences in data collection between studies (e.g., prospective vs retrospective, paper vs electronic reporting) and assignment of impact/severity are then further confounded by variations in reporting; of the 23 studies reviewed for this article (Table E1), only 12 (52%) reported the total number of doses administered, 15 (65%) the number of patients experiencing at least one AE during OIT, 10 (43%) the proportion of doses causing AEs, and 13 (57%) a severity grading for reactions. Reporting AEs using number of doses as the denominator gives the perception that AE are less common; in the STOP-2 study [27], there were 73 episodes of wheezing—reported as an AE in relation to a dose of OIT—in 21 patients. Is this best reported as 0.41% of doses administered, or as an AE which occurred in 22% of participants? Or should both figures be presented (as the study investigators chose to do), in order to address this issue?

Improving the safety of OIT

The need for consistency in AE collections and reporting is essential if we are going to be able to attempt even limited comparisons between different study protocols, with the aim to improve patient safety and efficacy outcomes. We have previously reviewed the options to improve the safety of OIT [8], and these can be summarised as strategies to:

  1. a)

    improve protocol designs—by use of hypoallergenic products (e.g., through food-processing or selection of specific peptides), modification of schedule intensity or the use of adjuvants such as omalizumab.

  2. b)

    improve patient selection, in order to offer individualised protocols which are better targeted to patients with “high-risk” profiles who are considered to be at greater risk of dose-related AEs and treatment failure.

This requires a two-pronged approach: one that aims to increase overall safety of OIT, and second, to specifically identify patients who might be at higher risk of AEs and treatment failure. Given finite resources, the focus should be on the predictability and prevention of (1) high-risk adverse events (anaphylaxis, recurrent gastrointestinal AEs) and (2) reactions during longer-term maintenance, where it is not uncommon for patients who have tolerated a daily dose for many weeks and even months to suddenly experience a breakthrough reaction including anaphylaxis.

Factors potentially contributing to dose-related AEs

These may be divided into three broad categories, as summarised in Fig. 1.

  1. 1.

    Protocol-related factors

    Several early studies used initial rush updosing phases which were associated with high AE rates [13]. However, this might have led to earlier treatment failures, with the consequence of selecting out higher-risk patients, thus reducing the overall dose-related AE rates during subsequent treatment phases. Many protocols now include a “rush” updosing phase over several days [28,29,30], although evidence that this is a more efficacious approach than slower, more extended initiation phases is lacking, as no head-to-head comparison has been undertaken.

    Dello Iacono et al. reported that frequency (but not severity) of AEs correlated with the size of increment at any given dose-step in an open-label trial of egg OIT [31]. The study team proposed that the higher rate of in-hospital AEs was a function of higher dosing increments in hospital compared with those undertaken at home. However, it is difficult to avoid confounding due to better reporting of in-hospital AEs compared with those at home, where minor symptoms might not always be recorded. In contrast, Vazquez-Ortiz et al. noted that for egg OIT, the greatest number of AEs occurred at the lowest and highest doses and did not necessarily relate to the dosing increment [32]. For peanut, Wasserman et al. did not observe any impact of dosing increment on risk of AEs [2].

    Multiple studies have reported higher rates of AEs during the initial updosing phase, which reduce over time [11, 20]. However, safety data relating to longer-term maintenance (where adherence may be reduced) are sparse. In the pooled analysis by Virkud et al. for peanut OIT, while overall rates for AEs were greater during the build-up phase, more patients received adrenaline (epinephrine) during maintenance phase (although the rate was not adjusted for number of doses in each phase and patient withdrawal); this might reflect increasing compliance with appropriate management of breakthrough reactions and/or reduced adherence to therapy [20].

    Finally, rates of AEs may also be related to the target maintenance dose. Paradoxically, in a comparison of high and low dose peanut OIT, Vickery et al. reported a significantly greater rate of moderate AEs in the lower dose group, which reversed during maintenance [11]. This suggests that a higher target might protect against AEs by inducing a faster and larger increase in the dose needed to trigger AEs, thus providing greater protection from dosing-related AEs. Although, it has been proposed that lower-dose protocols might be a useful strategy in desensitising higher-risk patients [33], it is difficult to make conclusions due to the lack of consistency in reported AE data between studies. No statistical difference was observed in rates of AEs in a recent report of lower and higher-dose OIT for wheat allergy [34].

  1. 2.

    Patient-related factors

  1. (a)

    Egg

Vazquez-Ortiz et al. reported that 90% of children (n = 50) receiving OIT to pasteurised raw egg experienced AEs [32]. Participants were divided into those with AEs that resolved over time (median time to resolution 7.6 months) (48%), those with persistent reactions (34%) and those who discontinued OIT due to side effects (18%). Participants with persistent reactions had more frequent and more severe reactions, while those who discontinued treatment had higher specific baseline IgE and a lower threshold at baseline challenge.

For both egg and milk OIT, Garcia-Lirio et al. noted the serum cytokine levels (IL-2, IL-4, IL-6, IL-10, IL-12, IL-17, IFNγ, and TNFα) did not change during OIT and therefore could not predict those patients who might be at greater risk of persistent reactions [35]. Unfortunately, the authors did not assess cytokine response of peripheral blood mononuclear cells following in vitro allergen stimulation.

  1. (b)

    Cow’s milk

In an open-label trial of milk-OIT, Salvilahti et al. reported a significant difference between baseline allergen-specific IgE and IgG4 levels, diversity and binding affinity (including the major component α-1-casein) in those successfully completing OIT compared with those withdrawing due to AEs [36]. Interestingly, participants who failed OIT had higher affinity for both IgE and IgG4 at baseline, while those in whom OIT was successful had greater overlap in IgE and IgG4 regions prior to OIT initiation, compared with treatment failures. Importantly, the diversity of binding did not significantly alter over time, and thus might provide a baseline indicator to help predict outcome. In another study, Martinez-Botas et al. found a linear association between the number of peptides demonstrating IgE binding and the number of AEs during milk-OIT [16].

Vazquez-Ortiz et al. undertook survival analysis in a cohort of 81 Spanish children undergoing cow’s milk OIT; the cumulative probability of resolution of dose-related AEs was 25% and 50% at 3 and 8 months respectively [37]. Allergen-specific IgE > 50 kua/L, skin prick wheal > 9 mm, and more severe reactions at pre-OIT food challenge were found to be independent risk factors for persisting AEs during OIT, with an additive/synergistic risk elevation on combination of risk factors.

Finally, Kuitunen et al. investigated predictors of treatment efficacy in a study of 76 children undergoing OIT to cow’s milk [38]. They found that participants who were classified as partial responders (tolerating < 200 ml cow’s milk at exit challenge) had a significantly higher proportion of adverse skin reactions during treatment, compared with those who tolerated > 200 ml following OIT (43% vs 18% of all adverse reactions, p = 0.038).

  1. (c)

    Peanut

Virkud et al. [20] undertook a pooled analysis of data from three studies of peanut OIT (two of which were not included in the meta-analysis by Nurmatov et al. [15]), with a total of 104 children included. Rates of systemic AEs increased 2.2 times (95% CI 1.1–4.3) in participants with allergic rhinitis, compared with those without. Patients with allergic rhinitis experienced dose-related AEs more frequently between April and October i.e., seasonality, while those without experienced more AEs between January and September in those without allergic rhinitis. However, whether this observation remained after correcting for date of OIT commencement and/or duration on OIT is unclear. A significant relation was noted between skin sensitisation and gastrointestinal AEs, with a 1.8-fold increase in gastrointestinal events for every 5 mm increase in baseline SPT size. However, 92% of this cohort were under 8 years old (55% under 3 years), so these findings need replication, both in the same and older age groups. Likewise, some of the included studies had a peanut-specific IgE ≥ 7 kUA/l as an inclusion criteria; as a result, only 12% of participants had a peanut-specific IgE < 7 kUA/L, and so the adverse event frequency might be higher than in the general peanut-allergic population.

Wasserman et al’s recent analysis of 270 patients undergoing office-based OIT reported that the odds of achieving successful desensitisation (defined as tolerance to a maintenance dose of 3 g peanut protein and “passing” a food challenge of 6 g peanut protein) were reduced with increasing age and higher baseline peanut-specific IgE [2]. These data are consistent with the conclusions of another study by Vickery et al. who undertook peanut OIT study in a younger cohort (aged 9 to 36 months) [11], though direct comparison is difficult due to differences in protocols and definitions of adverse reactions. As of the time of writing, no analysis of predictors of dose-related AEs has yet been published for the PALISADE study, currently the largest phase 3 randomised controlled study of peanut OIT to date [39].

  1. (d)

    Wheat

Kulmala recently published results from an open-label OIT study in 100 wheat-allergic children [17]. The authors report that baseline sensitisation to ω-5-gliadin but not whole wheat was associated with more severe AEs, although they note that some patients had very high baseline levels of antibodies ω-5-gliadin and yet had mild or even no reactions during the study. A smaller but randomised double-blind placebo-controlled study did not assess factors-associated AEs, but did report a significant rate of AEs in the placebo group (5.8% of all doses) compared with active treatment (15.4%) which reinforces the need to assess for placebo reactions when interpreting AE data [34].

  1. 3.

    Co-factors

    It is not uncommon for participants undergoing OIT to experience breakthrough reactions to a maintenance dose, despite having tolerated that dose for weeks or even months prior. Potential cofactors contributing to this can include intercurrent infections, exercise, tiredness, and menstruation [19], and it has been reported that most AEs during maintenance occurred in the context of cofactors [40]. However, it is difficult to determine the degree (if any) to which these factors contribute to the occurrence of the reaction, or whether their presence is just coincidental. There is a lack of data in relation to how frequently doses are tolerated despite the presence of these co-factors, which impedes these types of analysis. This remains a significant data gap, as it would be helpful to address, where there is a subset of patients who experience cofactor-related AEs and whom can be predicted in advance, thus reducing the need for dosing-related restrictions for other patients.

    Fig. 1
    figure 1

    Potential factors which may contribute to the occurrence and severity of dose-related adverse events during OIT

Can we identify higher-risk patients as a strategy to improve the safety of OIT?

Unfortunately, around 20% of patients fail OIT; these patients tend to experience a higher frequency of dose-related AEs but also share the characteristics associated with a tendency towards more severe reactions with accidental allergen exposure [18]. Given these are the patients most likely to benefit from OIT, there is a need to find strategies to improve the treatment safety profile through modified protocols [8, 41]. However, there are also some patients in whom the risk of current OIT protocols outweighs the risk associated with avoidance [42]. OIT is associated with more reactions than avoidance [43], although arguably for most patients undergoing OIT the increase in reaction frequency is still associated with an improvement in health-related quality of life.

Severe and life-threatening reactions during OIT have been reported [41, 42]. These patients are often adolescents (in whom compliance may be an issue) and have higher IgE levels and persistent asthma along with a history of recent missed doses or co-factors. However, these features cannot distinguish from many more children with the same characteristics, who safely achieve desensitisation. These and other patients with persistent food allergies may have specific epitope binding patterns, which could also signal a patient-group at higher risk of dose-related AEs and reduced efficacy during OIT, but this has only, as yet, been assessed for milk OIT [44].

Given the differences in protocols for published and ongoing OIT trials, one potential way forward would be to undertake sequential meta-analyses and grouped analyses. Meta-analyses have been published relating to OIT efficacy, but this has proven to be less feasible with respect to safety outcomes. A consensus would be needed to ensure alignment of key data outcomes and standardised reporting of OIT data. Such a consensus statement would require input from all stakeholders—most importantly, patients and their families—in order to consider the relevance of outcomes which bother patients most, rather than study investigators who may impose their own values assessment on the impact of AEs occurring in our patients. This would also provide an opportunity to better align AE terms with the medical dictionary for regulatory activities, which would be of use in moves towards introducing OIT more widely in clinical practice from the regulatory perspective [45]. A suggested minimum data set is proposed in Table 1.

Table 1 Suggested minimum reported of AEs in studies of OIT

Conclusions

The lack of consistency in dose-related AE reporting in studies of OIT has impeded our understanding of the safety of OIT, including the factors which might contribute to differences in safety outcomes and which strategies might help improve this, particularly for patients who are more likely to fail OIT or experience frequent AEs. Many of the OIT trials reported to date are limited by small cohort size, which makes generalisation of safety data to the wider food-allergic population difficult. Consensus is needed on AE reporting to allow pooled analyses which are likely to yield more useful predictors of OIT safety, and lead to improved protocols to maximise efficacy and safety to the benefit of our patients.