Introduction

Treating infections caused by gram-negative bacteria (GNB) resistant to multiple classes of antibiotics is an increasing challenge in many hospitals. Extended-spectrum β-lactamase (ESBL)-producing and carbapenem-resistant Enterobacteriaceae (CRE) in particular are now endemic in many parts of the world and represent one of the most serious public health threats [1, 2]. In this context, antimicrobial stewardship (AMS) programs play a critical role in improving appropriate antibiotic use to improve patient outcomes and reduce the antibiotic selective pressure that drives the emergence of further resistance. However, both the efficacy and collateral impact of many stewardship strategies are not well documented, starting from the battle against the overuse of carbapenems in non-critically ill patients to the use of carbapenem-sparing regimens also in critically ill patients in institutions with endemically high levels (>20%) of ESBL-producing Enterobacteriaceae.

Herein, we discuss the leading stewardship-related controversies and infection control (IC) measures needed to control infections caused by ESBL- and CR-producing Enterobacteriaceae. We also discuss recent efforts from our large teaching hospital to contain carbapenem resistance and its impact on other ESBL resistance rates. This article is based on previously conducted studies and does not involve any new studies of human or animal subjects performed by any of the authors.

ESBL-Producing Enterobacteriaceae

ESBL-producing GNBs are now a universal health concern [2]. ESBL production is frequently accompanied by other resistance mechanisms that impart cross-resistance to many other anti-GNB antibiotics, including aminoglycosides and fluoroquinolones. Consequently, carbapenems have long been regarded as the drugs of choice for suspected or proven serious infections caused by ESBL-producing organisms [36]. Nevertheless, the increased use of carbapenems driven by the spread of ESBL-producing Enterobacteriaceae has contributed to the emergence of carbapenem-resistant strains, which currently represent one of the main concerns in the management of patients with gram-negative infections [7, 8]. Avoidance of carbapenems for the treatment of ESBL-producing organisms is conceptually appealing from the standpoint of an AMS program, but could increase the risk of inappropriate treatment of relapse for patients with serious infections.

β-Lactam/β-lactamase inhibitor combinations (BLBLIs) such as piperacillin-tazobactam, have been recommended as an alternative to carbapenems for ESBL-producing Enterobacteriaceae if in vitro susceptibility is documented. Indeed, case-control studies from several countries have reported that frontline protocols that utilized piperacillin-tazobactam instead of cephalosporines or carbapenems were associated with a reduction in the risk of isolation of ESBL-producing and CRE, without clear increases in piperacillin-tazobactam resistance [8, 9].

However, there are some misgivings about the role of BLBLIs for the treatment of ESBL Enterobacteriaceae in clinical practice. First, BLBLIs may have diminished efficacy against ESBL-producing Enterobacteriaceae in the presence of a high bacterial load, known as the inoculum effect, which has been observed specifically with piperacillin-tazobactam [10]. Second, pharmacokinetic/pharmacodynamic (PK/PD) models may support the use of high doses of piperacillin/tazobactam for ESBL-producing Escherichia coli, but are much less encouraging for Klebsiella pneumoniae with higher minimum inhibitory concentrations (MICs) [11]. Third, increasing resistance to BLBLIs in ESBL producers (especially K. pneumoniae), overexpression of β-lactamases or complex coresistance mechanisms, including expression of other enzymes, such as plasmid-derived AmpC, may compromise the use of BLBLIs in empirical therapy.

Several cohort studies and a meta-analysis have suggested BLBLIs are safe and effective for ESBL-producing Enterobacteriaceae [1214], even though prospective clinical studies are scarce [15, 16]. The most relevant studies are two observational, well-designed studies on patients with bloodstream infections (BSIs) caused by ESBL-producing GNB, which reached opposite conclusions: according to Rodríguez-Baño et al. [12], BLBLIs were equivalent to carbapenems for treating ESBL-producing E. coli, while Tamma et al. [17] reported that empiric treatment with piperacillin-tazobactam versus carbapenem was associated with increased risk of mortality in BSI caused by ESBL-producing E.coli, Klebsiella spp. or Proteus spp. These contrasting findings are probably due to the differences in terms of etiological agents of BSI (E.coli only versus E.coli, Klebsiella spp. or Proteus spp.), administered dosages of piperacillin/tazobactam (higher doses were used in the study from Spain, 4.5 versus 3.375 g qid) and infection sources (>70% of BSIs in the Rodríguez-Baño et al. [12] study originated from the urinary tract versus much more variable infection sources in the study of Tamma et al. [17]). The impact of the infectious source on mortality has been highlighted by a further analysis of the Spanish cohort: all the patients with BSI arising from the urinary tract survived irrespective of the piperacillin-tazobactam MIC, while, in case of different BSI sources, the outcome was poorer if the piperacillin-tazobactam MIC was >2 mg/l [18].

Ideally, the role of BLBLIs versus carbapenems for the treatment of bloodstream infection caused by ESBL producers should be assessed in a large randomized controlled trial. Such a study is now recruiting patients (the ‘MERINO’ trial; NCT02176122) and aims to be completed by 2018.

Based on the data described to date, BLBLIs can represent a reasonable carbapenem-sparing treatment option for urinary tract infections, including those causing bacteremia, caused by ESBL-producing E. coli or, in case of non-severe infections, from other sites in which the isolate is susceptible at a low MIC (≤2 µg/ml in the case of piperacillin-tazobactam). If piperacillin-tazobactam is used, it should be administered at high doses and using infusion strategies that maximize the PK/PD parameters (i.e., t > MIC; loading dose followed by maintenance doses administered by extended or continuous infusion) to guarantee an adequate exposure [19].

Carbapenems would remain the drug of choice for non-urinary or non-easily drainable infections (e.g., pneumonia) caused by ESBL-producing bacteria with complex resistance patterns, including K. pneumoniae and multiple β-lactamases. However, de-escalation of carbapenems, recently described to be safe and feasible [20], probably now represents the most relevant and cost-effective intervention of any AMS program aimed to reduce carbapenem exposure. The actual “high-speed” microbiological turn around could be a cornerstone for these strategies, as further described.

The scenario described above will likely change in the future, owing to the introduction of novel BLBLIs in clinical practice [21]. Ceftazidime-avibactam, an established antipseudomonal cephalosporin combined with a new β-lactamase inhibitor, and ceftolozane (a new cephalosporine) combined with tazobactam have been approved by the Food and Drug Administration (FDA) for the treatment of complicated intraabdominal and urinary tract infections [2224]. These drugs could represent a further alternative to carbapenems to treat serious infections caused by ESBL-producing organisms, and this option for placement in therapy should be considered in the future hospital formularies. However, resistance to these newer cephalosporins or advanced spectrum β-lactamase inhibitors is already being reported among Enterobacteriaceae [25]; therefore, their judicious use is a key consideration for stewardship programs.

Carbapenem-Resistant Enterobacteriaceae

Over the last decade, the prevalence of infections with CRE has increased worldwide [1], and given the limited therapeutic options for these infections, they are associated with high morbidity and mortality. The crude mortality among patients with CRE infection is approximately 40% [26, 27]. The most relevant observational studies on risk factors for poor outcome in patients with CRE infection are summarized in Table 1. Predictors of poor outcome may be divided in three groups: (1) patient-related factors that included age, comorbidities and APACHE score; (2) infection-related factors such as the pulmonary source, BSI, severe sepsis or septic shock at infection onset, and isolation of a colistin-resistant strain; (3) treatment-related factors with prompt source control and combination therapy both associated with better outcome.

Table 1 Observational studies analyzing risk factors for poor outcome among patients with carbapenem-resistant Enterobacteriaceae infection

Given the limited number of effective and safe agents, several strategies have been proposed to treat CRE infections [2837]. The currently recommended strategy is combination antimicrobial regimens for CRE with the hope that synergistic interactions between agents will improve bactericidal activity, suppress the emergence of resistance and overcome the PK weaknesses of individual agents. Indeed, several studies have reported lower mortality rates (0–40%) among patients who received combination therapy versus patients receiving meropenem, colistin or tigecycline monotherapy (40–80%) [29, 33, 3638].

Unfortunately, the question of which combination is superior remains unresolved. In a meta-analysis of 12 retrospective cohort studies and case series, an advantage for combination therapy was identified only in patients classified as receiving “any combination” [39]. Among the different combinations, those that included high-dose carbapenems (i.e., 4–6 g meropenem administered by extended infusion) were associated with better outcome in some studies [29, 36, 37]. In the latest Italian study [36], a propensity score adjusted model to assess the impact on outcome of combination therapy revealed significantly improved survival among patients receiving a combination regimen that included a carbapenem. However, the protective role of carbapenems was maintained only for strains with MIC values ≤8 mg/l, representing around 30% of isolates in some settings with endemic carbapenem resistance.

The impact of carbapenem-sparing regimens on the outcome of CRE has only been investigated in a few small cohort studies. Combination regimens with high-dose tigecycline (100 mg every 12 h) have been proposed for patients with pulmonary infections and were associated with better outcomes than conventional doses of tigecycline in the treatment of ventilator-associated pneumonia caused by multidrug-resistant (MDR) GNB [30]. However, tigecycline use has been clearly associated with the emergence of reduced susceptibility to this drug [40]. In a recent report of 50 patients with highly resistant CR-K. pneumoniae (all strains resistant to colistin and with meropenem MIC >32 mg/l), regimens that included gentamicin, mostly in combination with other drugs, were the only combination associated with lower mortality [41].

The increasing prevalence of colistin-resistant strains, nearly 40% in some geographical areas, has required growing use of unconventional antibiotic combinations for CR-K. pneumoniae [42, 43]. Colistin plus rifampin, double carbapenem therapy and colistin plus double carbapenems have been proposed based on in vitro studies, animal models, and case reports or case series [28, 35, 44, 45]. However, the optimal regimen for such extremely drug-resistant/pan-drug-resistant (XDR/PDR) strains is still unknown.

Several new antibiotics, recently approved (i.e., ceftazidime-avibactam) or in development (RPX7009, plazomycin), have in vitro activity against CR-K. pneumoniae. Avibactam inhibits the activities of Ambler class A and C and some Ambler class D enzymes, restoring the activity of ceftazidime against ESBL, K. pneumoniae carbapenemases (KPC) and/or AmpC β-lactamases and P. aeruginosa. However, avibactam does not restore the spectrum of ceftazidime against metallo-β-lactamases (MBL), Acinetobacter baumannii and most gram-negative anaerobes. In a recent surveillance study of 124 CREs (KPC, n = 87; MBL, n = 13; OXA-48, n = 7), susceptibility rates to ceftazidime-avibactam for each group were 100, 0 and 85.7%, respectively [46]. Some authors also reported the compassionate use of ceftazidime-avibactam in four patients with infections due to Klebsiella oxytoca KPC (1) and K. pneumoniae OXA-48 (3); clinical cure was achieved in 2/4 (50%) [47].

Rapid Diagnostic Tests Used by AMS Programs

As previously affirmed, early discontinuation of broad-spectrum empirical antimicrobial regimens is one of the main goals of AMS programs. Tests for rapid identification of microorganisms represent a key tool for this purpose.

Several rapid molecular assays are commercially available [4850]. Peptide nucleic acid fluorescent in situ hybridization (PNA FISH; AdvanDx, Woburn, MA, USA), which uses synthetic oligonucleotide fluorescence-labeled probes that rapidly hybridize to species-specific ribosomal RNA, was one of the first. It can identify E. coli, P. aeruginosa and K. pneumoniae, but its inability to identify resistance genes in gram-negative organisms limits its usefulness. Indeed, no studies have evaluated the clinical impact of the use of this assay in patients with gram-negative infections.

Some clinical data are available about the implementation of another technique, matrix-assisted laser desorption/ionization time-of-flight mass spectrometry (MALDI-TOF MS), as an adjunctive tool to AMS programs. MALDI-TOF MS is a fast, reliable and cost-effective test using the technology of mass spectrometry (ionization and disintegration of a target molecule). The mass/charge ratio of the resulting molecular fragments is compared with those of well-characterized organisms in a database [51]. MALDI-TOF MS does not provide antimicrobial susceptibility data, but its integration in AMS programs is associated with a reduction in time to both active and optimal therapy, length of stay and hospital costs [5254]. Also in a challenging population constituted of 153 patients with BSIs caused by MDR and/or ESBL-producing GNB, integrating rapid diagnostics with AMS improved outcome. Time to effective and optimal antibiotic therapy was reduced from 89.7 to 32 h (P < 0.001) and from 80.9 to 23.2 h (P < 0.001), respectively. Total and intensive care unit length of stay decreased from 23.3 to 15.3 days (P = 0.0001) and from 16 to 10.7 days (P = 0.008), respectively. Mortality was lower (21% versus 8.9%, P = 0.01) during the intervention period, and the study intervention was identified as a predictor of survival [odds ratio (OR), 0.3; 95% confidence interval (CI) 0.12–0.79] at multivariate logistic regression [55]. Interestingly, the impact of MALDI-TOF MS appears to be higher when a stewardship team manages the results rather than when the same results are just provided to the patient’s attending physicians (reduced time to appropriate therapy in 28.8% and 44.6% of patients, according to the physician’s choices and stewardship team recommendations, respectively; P = 0.001) [56].

Multiplex polymerase chain reaction (PCR) can be used for simultaneous detection of multiple organisms and resistance markers. BioFire Diagnostics’ FilmArray blood culture identification (BCID) panel tests for 24 organisms, including Enterobacter cloacae complex, E. coli, Klebsiella species, Proteus species, Serratia marcescens, and antimicrobial resistance genes, including carbapenem resistance, are also used for detection.

Nanoparticle Probe Technology (Nucleic Acid Extraction and PCR Amplification) Nanosphere’s Verigene blood culture gram-negative (BC-GN) assay uses nucleic acid extraction and PCR amplification followed by hybridization of target DNA to capture oligonucleotides on a microarray. After hybridization, signal amplification of hybridized probes provides an automated qualitative analysis of results. This test identifies genus, species and genetic resistance determinants [KPC, New Deli metallo-β-lactamases (NDM), CTX-M, Verona integron-encoded metallo-β-lactamase (VIM), IMP and OXA β-lactamases genes] of the most common gram-negative organisms, with an approximate turnaround time of 2 h from blood culture positivity [57, 58]. Although there are no published clinical studies on this topic, potential advantages of the use of BC-GN in combination with an AMS approach have recently been described by Bork et al. [59]. In their interesting simulation, they evaluated 132 GN-BSI episodes, performed a theoretical evaluation of the time to effective and optimal antibiotic therapy based on BC-GN reporting and AMS team review (intervention) and compared them with actual antibiotic administration times obtained from chart review (controls). Effective and optimal antibiotic therapy could be achieved an average of 3.7 h (95% CI 1.3–6.2; P < 0.01) and 18.3 h earlier (95% CI 13.3–23.4; P < 0.01) in the intervention group than in the controls [59].

In settings with epidemic ESBL-producing and CR-producing Enterobacteriaceae, rapid microbiologic tests, in particular novel techniques that allow early identification of antimicrobial resistance patterns, can represent a key tool to improve patient outcomes and antimicrobial use. Microbiologic technologies must be integrated in AMS programs, as their clinical impact can be weak without a stewardship-driven educational intervention [56, 60].

Infection Control Challenges

In consideration of the steady increase of occurrence of infections caused by ESBL and carbapenemase-producing GNB, AMS programs have been incorporated into IC strategies designed to prevent the spread of MDR GNB. IC practitioners similarly support AMS efforts through monitoring of resistance trends and outbreaks in the institution, promoting compliance with hand hygiene, contact and transmission precautions, supporting education efforts and assessing adherence to care bundles. Effective prevention strategies can reduce the rate of hospital-acquired infections, decrease the use of additional antibiotics and, in turn, reduce the rate of MDR GNB.

Guidelines proposed by the European Society of Clinical Microbiology and Infectious Diseases (ESCMID) to prevent the diffusion of MDR GNB infection and colonization [61] identify four key aims for IC programs: (1) reduction of direct-contact transmission involving skin-to-skin contact and physical transfer of microorganisms to a susceptible host from an infected or colonized person via hand washing and wearing contact precautions; (2) early identification of colonized patients with active screening (culture) protocols at hospital admission followed by contact precautions and physical separation of colonized patients; (3) monitoring of cleaning performance and use of dedicated medical equipment for personal involved in the care of colonized or infected patients; (4) reduction in antimicrobial pressure favoring the selection of MDR GNB through AMS programs.

Similar measures are suggested for containing the spread of ESBL-producing strains in healthcare environments. However, data supporting the effectiveness of these recommendations, especially active screening protocols in institutions with endemically high resistance rates, are more controversial [61]. First, the epidemiology of ESBL-resistance is generally more complex than carbapenem-resistance in Enterobacteriacae. This complexity is compounded by the sometimes more sporadic nature of ESBL resistance. Each of these situations will need to be managed in different ways, depending on the risk to the patients involved.

When ESBL resistance is endemic to the institution, intensive IC efforts in the hospital alone and AMS may not be sufficient to reduce resistance. In contrast with other MDR bacteria, ESBL-mediated resistance is suspected to be primarily spread through the community and long-term care facilities and less so within within healthcare-related institutions [62]. Several reports have confirmed the environmental persistence of ESBL-producing E. coli in superficial water, wastewaters and farms [63, 64]. Antibiotic use in livestock as a growth-promoter appears to be a major contributing factor, yet controls on the prescription and use of antibiotics are almost nonexistent. Additionally, fecal carriage of ESBLs has been reported in healthy individuals residing in the community [65, 66]. As a result of this situation, a considerable proportion of patients infected with ESBL-producing strain in many European and US hospitals come from the community [67, 68].

Another controversial aspect of the circulation of the ESBL-producing strain is the method of transmission. The role of healthcare diffusion via the hands of healthcare workers during hospital outbreaks is predominant for K. pneumoniae strains but less clear for E. coli strains. In fact, several reports have shown the failure of contact precaution measures alone in preventing the diffusion of ESBL-producing E. coli during outbreaks [61, 69]. Moreover, in settings where contact precautions are not commonly used for containing the spread of ESBL strains, the rate of cross-transmission from healthcare workers' hands appeared to be low [70].

Several studies showed a possible spread of ESBL- and/or AmpC-producing bacterial strains transmitted via the food chain or food animal production environment [7173]. In this situation, it is reasonable to assume that a significant proportion of hospitalized patients may be colonized without detection before admission to the hospital and therefore are not assigned to isolation precautions. Further studies are needed to assess the cost effectiveness of a massive implementation of IC measures to contain the diffusion of ESBLs in endemic settings.

A clearer understanding of essential IC strategies has been described for limiting the spread of CRE. In a seminal study describing the containment of a country-wide outbreak of carbapenem-resistant K. pneumoniae in Israeli hospitals, Schwaber and colleagues [74] analyzed the impact of guidelines instituted by the Israeli Ministry of Health in 2007 that mandated physical separation and dedicated staffing for hospitalized carriers of CRE. Adherence to the guidelines was ensured by a dedicated task force that reviewed all aspects of IC and new CRE cases, and it intervened with additional measures when necessary. With the nationally implemented intervention, the rates of CRE acquisition were reduced by nearly fivefold, and the incidence of new infection relative to carrier prevalence declined significantly.

At an institutional level, Ciabotaro et al. [75] described the practical implementation and impact of the 2007 Israeli guidelines in their hospital. In the 40 months following implementation of the CRE patient screening, cohorting, cleaning, and education program, the investigators reported a 16-fold reduction in carbapenemase-resistant K. pneumoniae that was sustained for 30 months.

We recently performed a quasi-experimental study from 2010 to 2014 with 30-month follow-up of a multifaceted IC program combined with a stewardship initiative to reduce CRE BSIs in our 1420-bed university-affiliated hospital. The program, managed by a dedicated team, required CRE rectal culture screening for any patient transferred or admitted to high-risk units, cohorting or isolating positive patients, and using dedicated medical equipment and transport pathways within the hospital, education programs for patients, staff and caregivers, and an intensive hygiene program with a particular focus on hand washing and room cleaning. In terms of AMS, a coordinated program promoting carbapenem-sparing regimens was implemented to reduce selective antimicrobial pressure. Any prescription of the above-mentioned drugs was recorded by pharmacists, who notified the infectious disease consultants (IDCs) daily through an ad-hoc alert system. IDCs reviewed each prescription within 48 h with a bedside patient evaluation and recommended modifications to the prescribers whenever indicated.

Following the IC and AMS intervention, the incidence rate of CRE BSI (risk reduction 0.96, 0.92–0.99, P = 0.03) and CRE colonization (risk reduction 0.96, 0.95–0.97; P < 0.0001) decreased over 30 months (Fig. 1). Our analysis showed this reduction was not coincidental, as CRE-screening compliance rates were associated with the observed reduction in the incidence of CRE carriage in the hospitalized population, as detected by monthly rectal swab cultures. Moreover, when average compliance rates with the screening protocols waned after 20 months of the program (Fig. 1b), rates of CRE BSI and positive-rectal swabs rebounded and only decreased once average screening compliance rates had been restored above 75% [76].

Fig. 1
figure 1

Monthly incidence trends in KPC-carbapenemase producing K. pneumoniae bloodstream infections (a) and colonization (b) detected by rectal swab cultures following introduction of a comprehensive antimicrobial stewardship and infection control program to limit carbapenem resistance. Dotted lines represent the monthly incidence per 10,000 patient days. Solid lines represent the incidence trends. Bars in panel b indicate monthly compliance rates of the fraction of patients who were appropriately screened with rectal cultures for KPC-carbapenemase producing K. pneumoniae. IRR calculated by segmented Poisson regression. Despite declining CRE rates, the IRRs of ESBL-positive Enterobacteriaceae were stable during the study period (IRR 1.02, 95% CI 0.99–1.04; P = 0.06). Likewise, we did not observe marked changes in the rates of MDR among MDR non-fermenting GNB (IRR 1.04, 95% CI 0.22–1.22, P = 0.132) or vancomycin-resistant Enterococcus (IRR 0.98, 0.92–1.04; P = 0.57). IRR Incidence rate ratio, KPC K. pneumoniae carbapenemases

Support of the regional health authority [77] and our hospital administration as well as the strict application of the IC policy were key elements in the success of the program and the AMS initiatives. Interestingly, during the intervention period, we did not observe a significant decrease in the incidence of BSI due to ESBL-producing Enterobacteriaceae or other MDR GNR organisms, highlighting that in the absence of specific screening for ESBLs, strictly applied IC procedures for CRE are unlikely to blindly decrease other endemic resistance problems.

Decolonization Strategies for MDR GNB

Nearly 30 years of selective decontamination studies that included more than 50 randomized studies and 10 meta-analyses have suggested that, in general, selective oral decontamination (SOD) or selective digestive decontamination (SDD) has a favorable effect on mortality from serious infections in adult patients admitted to intensive care units (ICUs) [68, 78]. However, the largest studies were performed in ICUs with relatively low rates of antibiotic resistance in The Netherlands, and the efficacy of these approaches is less certain when used in patients who are already documented to be colonized with MDR organisms. Whether these results translate to settings with higher rates of endemic MDR GNB is debatable. Indeed, several observational studies [7982] and a small randomized trial [83] from ICUs with higher resistance rates have reported contradictory results regarding the benefits of SDD for decolonization of MDR GNB.

In the last few years, the efficacy and safety of SDD with non-absorbable antibiotics for the eradication of CRE carriage were evaluated in populations with various underlying conditions [82, 8487].

The use of oral regimens of colistin and/or gentamicin has been mainly proposed for some specific settings such as patients with hematological malignancies or candidates to major surgery including solid organ transplantation (SOT). However, there are several misgivings about the use of CRE decolonization. First, the need for decolonizing SOT candidates is under debate, mainly for liver transplantation; indeed, the risk for CRE infection has been shown to be similar between CRE colonized and noncolonized candidates [88, 89]. Second, the efficacy of decolonization was short term [8486]. Third, it has been associated with the emergence of further resistance [82]. In a large randomized controlled trial performed in The Netherlands, routine use of SDD protocols of colistin, tobramycin and amphotericin B prior to detected colonization with resistant organisms was associated with a 7% increase in the recovery of aminoglycoside-resistant organisms per month [90]. In a retrospective cohort study performed in surgical ICUs in Germany, Lubbert and colleagues [82] reported that gentamicin and colistin SDD achieved decolonization rates of KPC-2 carbapenemase-producing K. pneumoniae in 6/14 patients (43%), but was associated with a rapid increase in secondary bacterial resistance to both antibiotics. Selection of colistin-resistant strains in particular is a major concern because it may predispose patients to infections with pathogens that are untreatable with currently available antibiotics [91]. Therefore, the potential ecological impact of SDD and limited benefits suggest that SDD can only be used with careful microbiological monitoring for resistance development.

Clearly, more data are needed to define the role of preemptive versus targeted SOD and SDD in the management of MDR GNB and its potential collateral impact on resistance rates in the ICU or hospital.

Conclusions

AMS has been defined as “a marriage of IC and antimicrobial management finalized to share the principles of the optimized treatment between the bench to bedside point of view and the hospital-wide vision” [92]. According to this definition, in our opinion, the main aims of stewardship should be: promoting the appropriate use of antimicrobials to decrease the spread of infections caused by MDR organisms without forgetting the patient outcome.

There are two principal approaches: restrictive and persuasive. The oldest strategy is antibiotic restriction or pre-prescription authorization, which consists of the requirement for approval of the antibiotic from an infectious disease specialist. In endemic settings for ESBL infections, the restriction of some key antimicrobials may reduce the antibiotic pressure and rate of resistant pathogens. For example, a fluoroquinolone restriction strategy was demonstrated to increase fluoroquinolone susceptibility within a short span of 4 months and reduce the overall rate of ESBL and Clostridium difficile infection cases [93, 94].

The persuasive strategy usually consists of a post-prescription review by an infectious disease specialist of some of the antimicrobials prescribed by the attending physicians, providing advice if necessary. The advantage of this strategy may also comprise the improvement of the overall appropriateness of prescription and the opportunity for early de-escalation [95]. In a study evaluating the efficacy of an AMS program based on early carbapenem de-escalation, the patients enrolled in the program had a comparable clinical success, fewer adverse effects and a lower incidence of the development of resistance compared to controls [20].

Both of these AMS strategies have pros and cons; when compared in systematic reviews, the restrictive policy seems to allow more rapid results, mainly in terms of reducing antimicrobial consumptions, but these results are usually limited in the time because of the general low level of acceptance of this strategy by prescribers. The second one seems to be less effective in terms of reducing antimicrobial consumption, but has the advantage of improving the prescription appropriateness with a potential advantage on patient outcome.

Beyond the classical question about the best strategy between persuasive or restrictive actions, an important question remains unanswered: should stewardship programs be planned in a whole hospital vision or can we also think about setting-related, disease-related, drug-related projects? Another advocated limit of stewardship programs is the relative paucity, if not the absence, of multicenter studies across large healthcare systems; however, because the inter-center variability of clinical settings and missions is a distinctive feature of any healthcare system, it seems logical to ask ourselves which is the primary goal of any stewardship program: reproducibility or attainability in relation to the specific unmet needs of any single institution or organization?