FormalPara Take-home message

Dogma often dictates the management of acutely ill patients; however, the underpinning evidence base is often slight at best leading to potential over-interpretation and misapplication. We encourage critical reappraisal of current practices.

Introduction

Dogma represents an opinion believed to be true or irrevocable. It originates from the Greek word ‘δόγμα’ (‘that which one thinks is true’), a cousin of ‘δόξα’ (doxa) which itself derives from the verb δοκεῖν (dokein) meaning ‘to seem’, ‘to think’, ‘to accept’. Plato considered doxa a belief unrelated to reason and framed it as the opponent of knowledge [1]. Aristotle, however, took the view that doxa’s value was in its practicality and common usage and was the first step in finding knowledge [2]. ‘Fact’—something known or proved to be true—derives from the Latin factum, ‘do’. The original sense was an act but later it described a crime (as in “after the fact”). The current sense of ‘fact’, indicating truth or reality, dates from the late sixteenth century.

Fact and dogma are frequently superimposed in medicine. Everyday medical practice revolves around ‘facts’ enshrined in textbooks and management guidelines, and then reiterated in subsequent versions. However, such ‘facts’ often lack adequate source verification. Not infrequently, the source is absent, misinterpreted, misapplied from other situations, or outdated. Yet, ‘incestuous amplification’—a reinforcement of ideas and convictions held by like-minded individuals, and the wish to remain within a perceived tried-and-trusted comfort zone, tend to quash any iconoclastic challenge to these cherished beliefs.

Dogma plays an important part in current clinical practice as it offers long-established Aristotelian solutions that are considered to increase patient safety, optimise care, and standardise operating procedures. Strict adherence appeals as the dogma seems to work, or at least is assumed to do so. Treatment failures and covert side effects are conveniently ascribed to the patient, blaming frailty, comorbidity or illness severity rather than the intervention.

Reliance or, worse still, insistence, on dogma can deter clinicians from personalising care for the individual patient. The intervention may be appropriate for some but ineffective, wasteful or even harmful to others. It may suppress innovation for fear of error. Litigation may ensue as failure to follow an established dogma can be viewed as breach of duty, with arguments on causation often passing unchallenged for lack of appreciation of the weak or misquoted evidence base.

An excellent recent example of a successful challenge to dogma is the change in sepsis management guidelines that now recommend a more measured approach to commencing antibiotics [3,4,5]. In this article, we offer a further selection of long-standing management practices that are often taken as gospel but, in reality, the evidence base is non-existent, weak, outdated, and/or physiologically questionable. The practice is assumed to work when the patient improves or the prophylactically treated complication fails to materialise. Whether this is a consequence of the intervention itself, the body’s ability to withstand an unnecessary iatrogenic insult, or simply overkill is moot. We do not advocate wholesale change but hope that this article will remind the reader that all that glistens is not necessarily gold. We should also revisit long-established though scientifically dubious practices to confirm whether they remain valid in the current era, either in part or in totality.

Dogma: “Loop diuretics are needed to treat acute heart failure”

The recent European Society for Cardiology 2021 guideline describes intravenous diuretics as the ‘cornerstone of acute heart failure management’ albeit with a low (‘C grade’) level of evidence [6]. Clinical benefit is often ascribed to the diuresis that follows an intravenous furosemide bolus. The full urine bag is taken as proof positive of successful fluid removal from the lung that has led to symptomatic improvement. Yet this oft-stated paradigm can be readily challenged by conflicting observations. The diuretic effect only commences after 20–30 min and peaks after an hour or so, yet the patient has often improved markedly beforehand, an observation noted in 1966 [7]. The following year, significant clinical and biochemical improvement was reported within an hour of furosemide administration despite ongoing oliguria [8]. The lack of rapid radiological resolution of pulmonary oedema also argues against concurrent pulmonary water clearance. The X-ray does not ‘lag behind the patient’, as frequently cited, but simply demonstrates excess lung fluid still present despite symptomatic relief. It can take several weeks for clearance of pulmonary (interstitial and alveolar) oedema by lymphatic drainage and decreased ingress of fluid into the lung from raised hydrostatic pressures.

The benefit from intravenous furosemide arises from its rapid vasodilating effect [9]. Dikshit and colleagues described this change in cardiac loading conditions with a 27% fall in left ventricular filling pressure and a 52% increase in mean calf venous capacitance within 5–15 min following a furosemide bolus [10].

The downsides of furosemide should also be highlighted. Multiple studies report haemodynamic deterioration with falls in stroke volume and increases in heart rate, blood pressure and systemic vascular resistance [11,12,13]. This relates to arterial vasoconstriction induced by hypovolaemia and rapid activation of the neurohormonal axis [9, 13]. The corollary of hypovolaemia and compensatory rises in aldosterone and vasopressin is a decrease in renal blood flow, a fall in glomerular filtration and consequent oliguria. The knee-jerk response is to administer further, if not larger, doses of furosemide. Subsequent worsening in renal function is then blamed on the patient’s poor underlying cardiac status. A propensity-matched retrospective analysis in medical and cardiac patients admitted to the intensive care unit (ICU) and receiving loop diuretic use within 24 h of ICU admission reported an association between acute kidney injury and electrolyte abnormalities, though this did not impact on other outcomes [14].

The rationale for using furosemide as first-line therapy in patients with acute heart failure who are not intravascularly volume overloaded is thus questionable. Is it not better to aim for normovolaemia and optimal vasodilation, both arterial and venous, to optimise both ventricular filling and the resistance against which the dysfunctional ventricle has to pump? Nitrates are pharmacologically better suited to achieve rapid resolution of symptoms but with improved haemodynamics. At low dose, they are predominantly venodilators but an arterial dilating effect occurs with increasing dose. In addition, appropriate and cautious fluid loading may be advantageous to correct any hypovolaemic contribution to a compromised circulatory status which may be unmasked by vasodilator therapy. Few prospective randomised studies have compared nitrates and furosemide in acute heart failure. One open-label trial conducted in mobile coronary care units in Israel randomised 110 patients to high-dose furosemide plus low-dose isosorbide dinitrate, or the converse [15]. This design was at the insistence of the ethics committee since both agents were deemed essential. The rather large composite outcome of death, myocardial infarction and requirement for mechanical ventilation was significantly reduced (from 46 to 25%) in the high-dose nitrate/low-dose furosemide group. Smaller studies comparing furosemide against isosorbide dinitrate [16, 17] did not report outcomes but demonstrated a superior haemodynamic profile with nitrates. Similar findings have been made with sodium nitroprusside [18] and sodium nitrite [19]. An open-label randomised trial of sustained vasodilation using sublingual and transdermal nitrates and oral vasodilators against standard-of-care, however, reported no benefit with either all-cause mortality or rehospitalisation rates [20]. Notably, there was no difference in blood pressure and no sparing of furosemide use, with a fifth of patients in both groups suffering worse renal function. Another open-label randomised multicentre trial, performed in the emergency department, compared a bundle of care including intravenous nitrate boluses against standard-of-care in very elderly patients [21]. No outcome differences were shown though it is noteworthy that 98% of the intervention group received diuretics (median [IQR] dose 40 [40–80] mg). The study that needs to be done, in our view, is a comparison—preferably blinded—of diuretics against vasodilators with avoidance of diuretics unless specifically indicated (e.g. chronic diuretic use, true volume overload).

Dogma: “Every ICU patient should get heparin thromboprophylaxis”

Unless contraindicated by coagulopathy or active bleeding, the use of heparin has become a standard-of-care to prevent thromboembolic events in ICU patients. The underlying rationale is reasonable in that critically ill patients are generally immobile for prolonged periods and many inflammatory conditions are associated with a prothrombotic state [22]. Yet, how strong is the evidence that thromboprophylaxis actually makes a difference in this patient population?

In a 2013 meta-analysis, Alhazzani and colleagues could only identify three studies published between 1982 and 2000 comparing either unfractionated (UFH) or low molecular weight (LMWH) heparin prophylaxis against placebo in medical-surgical ICU patients [23]. Worryingly, the largest of the three trials, comprising 73% of included patients and which reported the most positive outcomes, has only been published in abstract form [24]. The other two studies did not record any cases of pulmonary embolism in their control groups. No effect was seen on symptomatic deep-vein thrombosis, mortality or bleeding rates.

For acutely ill medical patients (mainly non-ICU), a Cochrane meta-analysis [25] reported a significant reduction in deep venous thrombosis (DVT) incidence with heparin thromboprophylaxis (6.7%) versus either placebo or an unblinded control group (3.8%), but no significant impact on either non-fatal (0.5 vs. 0.2%) or fatal (0.3 vs. 0.2%) pulmonary embolism, nor on mortality. Major bleeding increased with heparin use (from 0.4 to 0.6%), although more so with unfractionated rather than LMW heparin.

The use of heparin thromboprophylaxis is more compelling in some surgical populations. In trauma patients, the heparin-treated group showed a significant reduction in DVT rates (8.7 to 4.3%) [26] and a non-significant reduction in pulmonary embolism rates (3.3 to 1.7%). Nonetheless, a meta-analysis of 3130 patients enrolled into eight randomised controlled trials (RCTs) for vascular surgery showed no significant difference in thromboembolic events nor bleeding complications [27]. A Cochrane Review including almost 3000 patients undergoing hip fracture repair reported a reduction in DVT events with any type of heparin, but no clear impact on fatal pulmonary embolisms or mortality, nor blood loss [28]. In patients with coronavirus disease 2019 (COVID-19) requiring intensive care, thromboembolic events, particularly pulmonary emboli, were fivefold higher compared to a matched non-COVID acute respiratory distress syndrome population [29]. Not unreasonably, heparin dosing was increased empirically to counter this risk. However, subsequent RCTs in ICU-level care patients showed no benefit of therapeutic dose over prophylactic (low or intermediate) dose heparin anticoagulation [30, 31], nor intermediate versus standard dose prophylaxis [32]. Of note, studies in non-ICU COVID patients reported a significant reduction in major thromboembolism events and death with therapeutic LMWH compared to low/intermediate dose LMWH anticoagulation [31, 33].

The other aspect of thromboprophylaxis that requires critical reappraisal is the evidence (or lack of) underpinning a fixed-dose subcutaneous LMWH regimen in ICU patients. This approach is advocated for ease of administration such that routine monitoring of anti-Factor Xa (anti-FXa) activity is not recommended. Nonetheless, what evidence does exist suggests many ICU patients are under-dosed, with pharmacodynamics in a critically ill population failing to mirror less sick ward cohorts. Priglinger et al. found significantly lower anti-FXa activity (e.g. 50% reduction in peak levels) after administration of enoxaparin 40 mg sc once daily in a critically ill ICU population with normal renal function compared to medical ward patients [34]. Robinson et al. also found 40 mg sc enoxaparin once daily was subtherapeutic in nearly 30% of their ICU patients, and 5% using 60 mg dosing [35]. In a mixed cohort of 219 ICU patients receiving enoxaparin thromboprophylaxis (40 mg sc non-obese, 60 mg obese, 20 mg renal failure), 30% were subtherapeutic in terms of anti-FXa activity and a further 30% just reached the therapeutic window [36]. In this study, the incidence of deep-vein thrombosis, adjudged by twice weekly ultrasound screening, was 2.7% yet all were asymptomatic. No pulmonary emboli were recognised. The question of once or twice daily dosing has also not been adequately addressed in the ICU setting [37].

Recent attention has been drawn to thromboprophylaxis in the COVID-19 population who are markedly pro-coagulopathic. Significantly lower mean anti-FXa activity was seen with 40 mg sc enoxaparin in COVID-19 patients in ICU patients compared to general ward patients (0.1 vs. 0.25 IU/ml), with 95% of ICU patients failing to achieve targeted anti-FXa activity levels [38]. Similar observations were made by Stattin et al. [39].

In summary, the evidence base supporting heparin thromboprophylaxis in all ICU patients is lacking in terms of outcome benefit. Worryingly, recommended fixed dosing regimens are often subtherapeutic. Nonetheless, the incidence of symptomatic DVT and pulmonary embolism is low. Whether, personalised targeted dosing does indeed provide an outcome difference, or whether alternatives such as direct oral anticoagulants (DOACs) are preferable, should be investigated in prospective studies.

Dogma: “A slow rate of sodium correction prevents central pontine myelinolysis”

Hyponatraemia (‘mild’ usually describes sodium values 130–135 mmol/l, ‘moderate’ between 120 and 130 mmol/l, and ‘severe’ < 120 mmol/l) is a common condition requiring hospitalisation. It is often due to excess sodium losses (including diuretic use) or excess water ingestion and, when lasting > 48 h, is deemed ‘chronic’ [40]. Risk of severe complications such as seizures and coma increases markedly with severe hyponatraemia. Management guidelines promote rapid correction by up to 5 mmol/l in the first hour for such severe complications, and thereafter gradual correction to avoid the feared complication of central pontine myelinolysis (perhaps more correctly termed osmotic demyelination syndrome as extrapontine myelinolysis is also reported). What constitutes ‘gradual’ remains a matter of conjecture, though more gradual correction is usually promoted for chronic hyponatraemia. National and expert panel recommendations are heterogenous. For instance, European guidelines [41] suggest targeting a rise of 5–10 mmol/l over the first 24 h, and then 8 mmol/l/24 h thereafter for both acute and chronic hyponatraemia. An American guideline from 2007 [42] recommended a rise < 10–12 mmol/l over 24 h, and < 18 mmol/l over 48 h for chronic hyponatraemia. No correction rate was offered for acute hyponatraemia. This was addressed in their revised 2013 guideline [43] where a review of the ‘limited available literature’ suggested an initial rapid rise of 4–6 mmol/l in serum sodium was sufficient to reverse the most serious manifestations of acute hyponatraemia; thereafter they considered that the correction rate need not be restricted for true acute hyponatraemia, nor was relowering of excessive correction indicated. A recent Best Practice document advised use of 3% hypertonic saline for any acute onset (< 48 h) and/or symptomatic hyponatraemia, and cited variable recommendations for correction rate ranging from 3–6 to 8–12 mmol/l/day in the first 24 h. For chronic asymptomatic hyponatraemia, they recommended a correction rate < 8 mmol/l/day [44].

This divergence of views largely relates to a lack of randomised, controlled trial data assessing correction rates. Three important questions to address are, first, the strength of evidence linking correction of hyponatremia to the development of osmotic demyelination syndrome; secondly, how commonly does osmotic demyelination syndrome arise; and, third, how often does osmotic demyelination syndrome cause permanent neurological disability.

A 1990 paper [45] performed magnetic resonance imaging in 13 patients with a mean sodium level at baseline of 103.7 mmol/l (range 93–113). The three patients who developed pontine lesions had a rise in mean (SD) serum sodium from 97.3 (6.7) to 127.3 (5.1) mmol/l in the first 24 h, i.e. a correction rate of 1.25 (0.4) mmol/l/hour. Of note, the ten patients not developing lesions still had a high correction rate of 0.74 (0.3) mmol/l/h, equating to 17.5 mmol/l over the first day. A 2015 systematic review characterised 158 cases with osmotic demyelination caused by hyponatraemia where sodium correction rates were described [46]. No difference in death nor degree of neurological outcome was seen, regardless of whether plasma sodium correction was slow (≤ 0.5 mmol/l/h) or rapid (> 0.5 mmol/l/h). Ayus and Arieff also failed to establish a correlation between patient outcome and either baseline level of sodium or the rate of correction [47]. In a retrospective analysis of 1490 patients with severe hyponatraemia (< 120 mmol/l), an overly rapid correction of serum sodium (defined as > 8 mmol in the first 24 h) occurred in 606 (41%) cases [48]. However, osmotic demyelination confirmed by MRI was described in only nine patients (0.6%) of whom one presented with demyelination at admission. Of the eight who developed subsequent demyelination, four had correction rates falling within the 10–12 mmol/l threshold recommended by the US [43] and European [41] guidelines.

Nearly, a quarter of patients who develop osmotic demyelination do so irrespective of hyponatraemia [49]. Other risk factors identified were alcohol abuse (54.7% of cases), cirrhosis (17.2%), malnutrition (16.4%) and liver transplantation (12.9%). Of the eight patients cited above [48] who developed osmotic demyelination, five had beer potomania, four had chronic alcohol abuse, and four were malnourished.

In terms of recovery, most studies report good functional outcomes either in the short or long term. A recent multicentre randomised trial of 178 patients with symptomatic hyponatraemia (sodium ≤ 125 mmol/l, 111 (62.5%) of whom had chronic hyponatraemia) compared rapid intermittent bolus vs slow continuous infusion of hypertonic saline [50]. Both strategies proved effective and safe though, even in this trial setting, 20.8% suffered from over-correction (increase in serum sodium > 12 mmol/l in 24 h or 18 mmol/l within 48 h). No patient, however, developed osmotic demyelination syndrome.

In summary, overly aggressive correction can predispose to osmotic demyelination [45]. However, all recent studies with more moderate correction rates (e.g. below 0.75 mmol/l/h) do not show any relationship with poor outcomes. Furthermore, osmotic demyelination is an uncommon event. A slower correction rate should perhaps be applied if other predisposing risk factors exist such as alcoholism and malnutrition. We should acknowledge the difficulty in precise hour-by-hour correction of sodium levels due to intrinsic patient variability and the challenge of balancing sodium and fluid input against losses (diuresis, vomiting, and diarrhoea).

Dogma: “Each hour counts for treating meningitis”

For acute bacterial meningitis, the current paradigm is to administer antibiotics as soon as possible, even commencing in the community. This dogma is applied to meningitis more than any other infectious condition. While we do not disagree with treatment urgency, the literature does not reflect this long-standing belief that “each hour counts”. Not surprisingly, there are no randomised controlled trials but 16 retrospective observational studies have sought an association between time-to-antibiotic from presentation (usually at hospital admission) to mortality (Supplement Table 1). Only one study [51] reported on duration of symptoms and trajectory of deterioration prior to admission and few reported on the incidence of long-term neurological sequelae. Overall, worse outcomes were associated with treatment delays post-hospital admission exceeding > 4–6 h, usually related to delayed diagnosis due to non-classical presentations. Examples are shown in Fig. 1a [52] and 1b [53]. Two studies reported an increasing risk of an unfavourable outcome per hour delay. Koster-Rasmussen [54] gave an odds ratio of 1.09 (95% confidence interval 1.01–19; p = 0.035) per hour delay for an unfavourable outcome, but did not report separately on mortality. Paradoxically, the effect was greater (OR 1.30 per hour, 95% CI 1.08–1.57, p < 0.01) when they excluded patients with a delay in antibiotic administration greater than 12 h. Glimaker et al. indicated a relative mortality increase of 8.8% (95% CI 3.4–14.4%; p < 0.01) per hour of delay [55]. However, the overall mortality rate was only 9.6% and so this equates to an absolute mortality increase of less than 1% per hour of delay. They provided a linear relationship extending to 14 h of delay; however, the mortality rate was curvilinear with little change over the first 4 h and with very wide and overlapping 95% confidence intervals (Fig. 1c). They reported that the risk of neurological sequelae was significantly associated with gender, age, aetiology, and mental status on admission, but made no mention of antibiotic delay.

Fig. 1
figure 1

Mortality related to time-to-antibiotics in three studies. A Case fatality rate according to door-to-antibiotic time interval in adult acute bacterial meningitis. (Figure redrawn from Ref [52]. By permission of Oxford University Press). B Time to antibiotic therapy and in-hospital mortality in community-acquired bacterial meningitis and time-to-antibiotic therapy and unfavourable outcome at discharge. *P value < 0.05 compared with patients treated 0–2 h from admission. (Figure redrawn from Ref [53]. By permission of Creative Commons Attribution 4.0 International License http://creativecommons.org/licenses/by/4.0/). C Probability of death related to time from admission to start of antibiotic treatment with 95% confidence intervals. (Figure redrawn from Ref [55]. By permission of Oxford University Press)

An interesting statistical commentary by Perera [56] highlighted Simpson’s paradox when performing analyses on such data, whereby the direction of the estimated effect can even change from benefit to harm. He noted that adjustment must be made for disease severity at the point at which the decision is made to give antibiotics as disease progression and the lack of specific signs and symptoms early in the illness are two important sources of confounding. Unfortunately, such data are not generally available in retrospective database analyses though this is clearly demonstrated by Bodilsen et al. (Fig. 1b) where patients treated with prehospital antibiotics had a higher mortality compared to patients treated post-admission. Such confounding has been recognised in sepsis; patients presenting in shock are more likely to be identified sooner and aggressively treated though their baseline mortality risk will be much higher [57]. Conversely, prolonged treatment delay is more likely in patients presenting with vague, non-specific features [58]. After adjustment for this particular confounder, time-to-antibiotic was not associated with mortality (adjusted OR 1.01; 95% CI 0.94–1.08; p = 0.78).

Age, mental status at admission and systemic disturbance prognosticate for poor outcomes [51, 55, 59, 60]. Aronin et al. [60] developed a prognostic model with low-, intermediate- and high-risk subgroups based on the absence (low-risk) or presence of hypotension, altered mental state and seizures either alone (intermediate-risk) or in combination (high-risk). Adverse clinical outcomes (death or persisting neurologic deficit) were more frequent in 42 patients progressing from low/intermediate to high risk before receiving antibiotics. No benefit was seen in 194 patients who remained at the same prognostic stage.

Delayed presentation in the community is also a poor prognosticator [51]. Of 286 patients with community-acquired bacterial meningitis, 125 had unfavourable outcomes. Pre-hospital delay in starting antibiotics from either onset of disease or onset of altered conscious level was a median one day longer in those suffering unfavourable outcomes. However, antibiotic delay post-admission did not impact on outcomes.

Finally, the long-running debate about delaying antibiotic administration until after computed tomography (CT) scan and lumbar puncture have been performed also remains unresolved. Glimaker et al. [55] demonstrated safety and outcome improvements with antibiotics given before lumbar puncture but pre-CT scanning, even in patients with moderate-severe impairment of mental status and/or new onset seizures. On the other hand, in their cohort of 1536 patients, Costerus et al. found no outcome impact in delaying antibiotics until post-CT scan [61].

In summary, in terms of outcome improvement, the urgency in treating meningitis appears to relate principally to severity of illness on presentation and a rapid trajectory of deterioration. This reflects findings for sepsis in general [3,4,5].

Conclusions

As we illustrate in the above examples, pros and cons of which are summarised in Fig. 2, adherence to dogma is commonplace in critical care practice. We could have offered multiple other examples where either the physiological rationale is questionable or the evidence base does not support either the practice nor the rigidly held beliefs. For lack of space in this article, we have not performed a formal systematic review on each dogma but believe we have captured the most important publications, utilising both search engines and reference lists from recent papers. In addition, we have scrutinised the content of each paper in depth and not relied on the headline findings in the abstract. We fully acknowledge that our interpretations of the data could potentially be challenged but this will lead to healthy, open discussion.

Fig. 2
figure 2

Pros and cons related to each presented dogma

Clearly, there are many instances where the principle underlying the dogma must be accepted. No one would argue against treating symptomatic hypoglycaemia or life-threatening hyperkalaemia but specific aspects could be challenged. For example, what level of acute hyperkalaemia is life threatening? What protection, if any and for how long, is offered by an intravenous injection of calcium given as a membrane stabiliser? The general perception is that calcium is a non-harmful intervention (or the possibility of harm is not even considered), but can we be sure?

It is easier to challenge established practice when the condition is frequently seen and high-quality evidence can be amassed. Updated iterations of cardiac arrest guidelines have reflected this increasing knowledge base though, notably, some components have been discarded for lack of proof of benefit, e.g. the use of atropine for asystole, or bicarbonate to reverse acidosis. Amassing evidence is more problematic in less commonly seen situations so there is an understandable reliance on ‘tried and trusted’ protocols, aspects of which may be superfluous or even harmful.

How do we move forward? As a good starting point, we should remind each other, and particularly trainees, that protocols and policies should not be blindly accepted. Expert opinion carries a not insignificant risk of academic bias. As part of medical training, statements and positions must not be taken at face value but the underpinning literature carefully scrutinised for balance, misinterpretation and important omissions. The above-cited studies highlighting frequent subtherapeutic dosing of fixed LMWH regimens in ICU patients are a good example.

Big data drawn from multiple centres could be applied to examining low-incidence conditions, comparing outcomes and complications where different interventions are applied. ‘Nudges’ could be deployed by computerised prescribing systems to propose suggested interventions where equipoise or uncertainty exists about specific treatments [62].

Fear of litigation can be assuaged by hospitals reclassifying their clinical management policies as guidelines or recommendations. Policies are interpreted by prosecution lawyers as rules of stone that do not allow practice deviation. Breach of duty is often claimed for over-rapid correction of hyponatraemia but, as illustrated above, the literature is not supportive of causation unless correction is excessive. Medicine is not one-size-fits-all and the clinician should be allowed to personalise care, albeit providing contemporaneous justification in the patient records for selecting a particular management strategy (Fig. 3).

Fig. 3
figure 3

Possible solutions to move forward