1 Introduction

Supersymmetry (SUSY) [1,2,3,4,5,6] introduces a symmetry between fermions and bosons, resulting in a SUSY partner (sparticle) for each Standard Model (SM) particle, with identical mass and quantum numbers, and a difference of half a unit of spin. Squarks (\(\tilde{q}\)), gluinos (\(\tilde{g}\)), sleptons (\(\tilde{\ell }\)) and sneutrinos (\(\tilde{\nu }\)) are the superpartners of the quarks, gluons, charged leptons and neutrinos, respectively. The SUSY partners of the gauge and Higgs bosons are called gauginos and higgsinos, respectively. The charged electroweak gaugino and higgsino states mix to form charginos (\(\tilde{\chi }_{i}^{\pm }\), \(i = 1,2\)), and the neutral states mix to form neutralinos (\(\tilde{\chi }_{j}^{0}\), \(j = 1,2,3,4\)). Finally, the gravitino (\(\tilde{G}\)) is the SUSY partner of the graviton. As no supersymmetric particle has been observed, SUSY must be a broken symmetry. In general, the minimal supersymmetric Standard Model (MSSM) allows couplings which violate baryon- and lepton-number conservation, leading to, for example, a short proton lifetime. To ensure accordance with established measurements, R-parity [7,8,9,10,11] conservation is often assumed. In this scenario, sparticles are produced in pairs and decay through cascades involving SM particles and other sparticles until the lightest sparticle (LSP), which is stable, is produced.

Final states with tau leptons are of particular interest in SUSY searches, although they are experimentally challenging. Light sleptons could play a role in the co-annihilation of neutralinos in the early universe, and models with light scalar taus are consistent with dark-matter searches [12]. Furthermore, should SUSY or any other physics beyond the Standard Model (BSM) be discovered, independent studies of all three lepton flavours are necessary to investigate the coupling structure of the new physics, especially with regard to lepton universality. If squarks and gluinos are within the reach of the Large Hadron Collider (LHC), their production may be among the dominant SUSY processes.

This article reports on an inclusive search for squarks and gluinos produced via the strong interaction in events with jets, at least one hadronically decaying tau lepton, and large missing transverse momentum from undetected LSPs. Two distinct topologies are studied, with one tau lepton (\(1\tau \)) or two or more tau leptons (\(2\tau \)) in the final state. These mutually exclusive channels are optimised separately. The analysis is performed using 3.2 \({\mathrm{{fb}^{-1}}}\) of proton–proton (pp) collision data at \(\sqrt{s}=13 {\mathrm {\ TeV}}\) recorded with the ATLAS detector at the LHC in 2015. Two SUSY scenarios are considered: a gauge-mediated SUSY-breaking (GMSB) model, and a simplified model of gluino pair production. Previous searches based on the same final state have been reported by the ATLAS [13, 14] and CMS [15] collaborations.

In GMSB models [16,17,18], SUSY breaking is communicated from a hidden sector to the visible sector by a set of messenger fields that share the gauge interactions of the SM. SUSY is spontaneously broken in the messenger sector, leading to massive, non-degenerate messenger fields. Gauginos acquire their masses via one-loop diagrams involving messengers. Squarks and sleptons, which do not couple directly to the messenger sector, get their masses from two-loop diagrams involving messengers and SM gauge bosons, or messengers and gauginos. The free parameters of GMSB models are the SUSY-breaking mass scale in the messenger sector (\(\Lambda \)), the messenger mass scale (\(M_\mathrm {mes}\)), the number of messenger multiplets (\(N_5\)) of the \(\mathbf {5}+{\varvec{\bar{5}}}\) representation of \(\text {SU}(5)\), the ratio of the two Higgs-doublet vacuum expectation values at the electroweak scale (\(\tan \beta \)), the sign of the Higgsino mass term in the superpotential (\(\mathrm { sign}(\mu )=\pm 1\)), and a gravitino-mass scale factor (\(C_\mathrm {grav}\)). Masses of gauginos and sfermions are proportional to \(\Lambda \), and scale as \(N_5\) and \(\sqrt{N_5}\), respectively. The \(M_\mathrm {mes}\) scale is required to be larger than \(\Lambda \), to avoid tachyonic messengers and charge- and colour-breaking vacua, and lower than the Planck mass to suppress flavour violation. The latter condition implies that the lightest supersymmetric particle is a very light gravitino. The \(C_\mathrm {grav}\) parameter, which results from the mechanism communicating SUSY breaking to the messengers, mainly affects the decay rate of the next-to-lightest supersymmetric particle (NLSP) into the LSP.

As in previous ATLAS searches [13, 14], the GMSB model is probed as a function of \(\Lambda \) and \(\tan \beta \), and the other parameters are set to \(M_\mathrm {mes} =250\) TeV, \(N_{\mathrm 5}=3\), \(\mathrm {sign}(\mu )=1\) and \(C_\mathrm {grav}=1\). For this choice of parameters, the NLSP is the lightest scalar tau (\(\tilde{\tau }_1 \)) for large values of \(\tan \beta \), while for lower \(\tan \beta \) values, the \(\tilde{\tau }_1 \) and the superpartners of the right-handed electron and muon (\(\tilde{e} _{\text {R}},\tilde{\mu } _{\text {R}}\)) are almost degenerate in mass. The squark–antisquark production mechanism dominates at high values of \(\Lambda \). A typical GMSB signal process is displayed in Fig. 1a. The value of \(C_\mathrm {grav}\) corresponds to prompt decays of the NLSP. The region of small \(\Lambda \) and large \(\tan \beta \) is unphysical since it leads to tachyonic states.

Fig. 1
figure 1

Diagrams illustrating the two SUSY scenarios studied in this analysis. In the GMSB model, the scalar lepton \(\tilde{\ell }\) is preferentially a scalar tau \(\tilde{\tau } _1\) for high values of \(\tan \beta \)

The other signal model studied in this analysis is a simplified model of gluino pair production [19] in an R-parity-conserving scenario. It is inspired by generic models such as the phenomenological MSSM [20, 21] with dominant gluino pair production, light \(\tilde{\tau }_1\) and a \(\tilde{\chi }_1^0\) LSP. Gluinos are assumed to undergo a two-step cascade decay leading to tau-rich final states, as shown in Fig. 1b. The two free parameters of the model are the masses of the gluino (\(m_{\tilde{g}}\)) and the LSP (\(m_{\tilde{\chi }_1^0}\)). Assumptions are made about the masses of other sparticles, namely the \(\tilde{\tau }_1\) and \(\tilde{\nu }_\tau \) are mass-degenerate, and the \(\tilde{\chi }_2^0\) and \(\tilde{\chi }_1^\pm \) are also mass-degenerate, with

$$\begin{aligned} m_{\tilde{\chi }_1^\pm } = m_{\tilde{\chi }_2^0}= \frac{1}{2}(m_{\tilde{g}} + m_{\tilde{\chi }_1^0}), \quad m_{\tilde{\tau }_1}=m_{\tilde{\nu }_\tau }=\frac{1}{2}(m_{\tilde{\chi }_1^\pm } + m_{\tilde{\chi }_1^0}). \end{aligned}$$
(1)

Gluinos are assumed to decay to \(\tilde{\chi }^{\pm }_{1} q \bar{q}'\) and \(\tilde{\chi }^0_{2} q \bar{q}\) with equal branching ratios, where \(q,q'\) denote generic first- and second-generation quarks. Neutralinos \(\tilde{\chi }^0_{2}\) are assumed to decay to \( \tilde{\tau } \tau \) and \( \tilde{\nu }_{\tau } \nu _{\tau }\) with equal probability, and charginos \(\tilde{\chi }^{\pm }_{1}\) are assumed to decay to \( \tilde{\nu }_{\tau }\tau \) and \( \tilde{\tau } \nu _{\tau }\) with equal probability. In the last step of the decay chain, \(\tilde{\tau } \) and \(\tilde{\nu }_{\tau }\) are assumed to decay to \(\tau \tilde{\chi }^0_{1}\) and \( \nu _\tau \tilde{\chi }^0_{1}\), respectively. All other SUSY particles are kinematically decoupled. The topology of signal events depends on the mass splitting between the gluino and the LSP. The sparticle decay widths are assumed to be small compared to sparticle masses, such that they play no role in the kinematics.

2 The ATLAS detector

The ATLAS experiment is described in detail in Ref. [22]. It is a multi-purpose detector with a forward–backward symmetric cylindrical geometry and a solid angleFootnote 1 coverage of nearly 4\(\pi \).

The inner tracking detector (ID), covering the region \(|\eta |<2.5\), consists of a silicon pixel detector, a silicon microstrip detector and a transition radiation tracker. The innermost layer of the pixel detector, the insertable B-layer [23], was installed between Run 1 and Run 2 of the LHC. The inner detector is surrounded by a thin superconducting solenoid providing a 2T magnetic field, and by a finely segmented lead/liquid-argon (LAr) electromagnetic calorimeter covering the region \(|\eta |<3.2\). A steel/scintillator-tile hadronic calorimeter provides coverage in the central region \(|\eta |<1.7\). The end-cap and forward regions, covering the pseudorapidity range \(1.5<|\eta |<4.9\), are instrumented with electromagnetic and hadronic LAr calorimeters, with either steel, copper or tungsten as the absorber material.

A muon spectrometer system incorporating large superconducting toroidal air-core magnets surrounds the calorimeters. Three layers of precision wire chambers provide muon tracking coverage in the range \(|\eta |<2.7\), while dedicated fast chambers are used for triggering in the region \(|\eta |<2.4\).

The trigger system, composed of two stages, was upgraded [24] before Run 2. The Level-1 trigger system, implemented with custom hardware, uses information from calorimeters and muon chambers to reduce the event rate from 40 MHz to a maximum of 100 kHz. The second stage, called the High-Level Trigger (HLT), reduces the data acquisition rate to about 1 kHz. The HLT is based on software and runs reconstruction algorithms similar to those used in the offline reconstruction.

3 Data and simulation samples

The data used in this analysis consist of pp collisions at a centre-of-mass energy of \(\sqrt{s}=13\) TeV delivered by the LHC with a 25 ns bunch spacing and recorded by the ATLAS detector from August to November 2015. Data quality requirements are applied to ensure that all sub-detectors were operating normally, and that LHC beams were in stable-collision mode. The integrated luminosity of the resulting data set is \(3.16 \pm 0.07\) \({\mathrm{{fb}^{-1}}}\).

Simulated Monte Carlo (MC) event samples are used to model both the SUSY signals and SM backgrounds, except multi-jet production, which is evaluated from data. All MC samples are generated at \(\sqrt{s}=13\) TeV. In addition to the hard-scattering process, soft pp interactions (pile-up) are included in the simulation using the Pythia 8.186 [25] generator with the A2 [26] set of tuned parameters (tune) and MSTW2008LO [27] parton density function (PDF) set. Generated events are reweighted such that the average number of pp interactions per bunch crossing has the same distribution in data and simulation. For SM background samples, the interactions between generated particles and the detector material are simulated [28] using Geant4 [29] and a detailed description of the ATLAS detector. In the case of signal samples, a parameterised fast simulation [30] is used to describe the energy deposits in the calorimeters.

The W+jets and Z+jets processes are simulated with the Sherpa 2.1.1 [31] generator. Matrix elements (ME) are calculated for up to two partons at next-to-leading order (NLO) and up to four additional partons at leading order (LO) in perturbative QCD using the OpenLoops [32] and Comix [33] matrix element generators, respectively. The polarisation of tau leptons in \(W(\tau \nu )\)+jets and \(Z(\tau \tau )\)+jets events is handled by the TauSpinner [34] program. The phase-space merging between the Sherpa parton shower (PS) [35] and matrix elements follows the ME+PS@NLO prescription [36]. The CT10 [37] PDF set is used in conjunction with dedicated parton-shower tuning. Simulated samples are generated in bins of the transverse momentum (\(p_{\text {T}}\)) of the vector boson. The inclusive cross sections are normalised to a next-to-next-to-leading-order (NNLO) calculation [38] in perturbative QCD based on the FEWZ program [39].

For the simulation of \(t\bar{t}\) and single-top-quark production in the Wt- and s-channels, the Powheg-Box v2 [40] generator is used with the CT10 PDF set for the matrix elements calculation. Electroweak t-channel single-top-quark events are generated using the Powheg-Box v1 generator. This generator uses the four-flavour scheme for the NLO matrix element calculation together with the fixed four-flavour CT10f4 PDF set. For all top quark processes, top quark spin correlations are taken into account (for t-channel production, top quarks are decayed using MadSpin [41]). The parton shower, hadronisation, and underlying event are simulated using Pythia 6.428 [42] with the CTEQ6L1 [43] PDF set and the corresponding Perugia 2012 tune [44]. Cross sections are calculated at NNLO in perturbative QCD with resummation of next-to-next-to-leading logarithmic (NNLL) soft gluon terms using the Top++ 2.0 program [45].

Diboson production is simulated using the Sherpa 2.1.1 generator with the CT10 PDF set. Processes with fully leptonic final states are calculated with up to one (\(4\ell \), \(2\ell +2\nu \)) or no partons (\(3\ell +1\nu \)) at NLO and up to three additional partons at LO. Diboson processes with one of the bosons decaying hadronically and the other leptonically are simulated with up to 1 (ZZ) or 0 (WW, WZ) partons at NLO and up to 3 additional partons at LO. The generator cross sections are used for these samples.

The simplified-model signal samples are generated using MG5_aMC@NLO v2.2.3 [46] interfaced to Pythia 8.186 with the A14 tune [47] for the modelling of the parton shower, hadronisation and underlying event. The ME calculation is performed at tree level and includes the emission of up to two additional partons. The PDF set used for the generation is NNPDF23LO [48]. The ME–PS matching is done using the CKKW-L prescription [49], with a matching scale set to one quarter of the gluino mass. The GMSB signal samples are generated with the Herwig++ 2.7.1 [50] generator, with CTEQ6L1 PDFs and the UE-EE-5-CTEQ6L1 tune [51], using input files generated in the SLHA format with the SPheno v3.1.12 [52] program. The parton shower evolution is performed using an algorithm described in Refs. [50, 53,54,55]. Signal cross sections are calculated at NLO in the strong coupling constant, adding the resummation of soft gluon emission at next-to-leading-logarithm accuracy [56,57,58,59,60]. The nominal cross section and the uncertainty are taken from an envelope of cross-section predictions using different PDF sets and factorisation and renormalisation scales, as described in Ref. [61].

4 Reconstruction of final-state objects

This analysis primarily requires the presence of jets, hadronically decaying tau leptons and missing transverse momentum in the final state. Jet b-tagging is used to separate the top quark background from vector bosons produced in association with jets (V+jets, where \(V=W,Z\)). Electrons and muons are vetoed in the \(1\tau \) channel, and muons are explicitly used in the \(2\tau \) channel for background modelling studies.

Primary vertices are reconstructed using inner-detector tracks with \(p_{\text {T}} >400 {\mathrm {\ MeV}}\) that satisfy requirements on the number of hits in silicon tracking devices [62]. Primary vertex candidates are required to have at least two associated tracks, and the candidate with the largest \(\sum p_{\text {T}} ^2\) is chosen as the primary vertex.

Jets are reconstructed with the anti-\(k_t\) clustering algorithm [63] with a distance parameter \(R=0.4\). Clusters of topologically connected calorimeter cells [64] with energy above the noise threshold, calibrated at the electromagnetic energy scale, are used as input. The jet energy is calibrated using a set of global sequential calibrations [65, 66]. Energy from pile-up interactions is subtracted based on the jet area and the median energy density computed for each event [67]. Jets are required to have \(p_{\text {T}} > 20 {\mathrm {\ GeV}}\) and \(|\eta | < 2.8\). The discrimination between hard-interaction jets and pile-up jets is achieved by a jet-vertex-tagging algorithm [68]. Events with jets originating from cosmic rays, beam background or detector noise are vetoed using the loose quality requirements defined in Ref. [69]. Jets containing b-hadrons (b-jets) are identified using a multivariate algorithm exploiting the long lifetime, high decay multiplicity, hard fragmentation, and large mass of b-hadrons [70]. The b-tagging algorithm identifies genuine b-jets with an efficiency of approximately 70% in simulated \(t\bar{t}\) events. The rejection rates for c-jets, hadronically decaying tau leptons, and light-quark or gluon jets are approximately 8, 26 and 440, respectively [71].

Electron candidates are reconstructed from an isolated energy deposit in the electromagnetic calorimeter matched to an inner-detector track. They are required to have \(p_{\text {T}} > 10 {\mathrm {\ GeV}}\), \(|\eta | < 2.47\), and candidates reconstructed in the transition region between the barrel and end-cap calorimeters (\(1.37<|\eta |<1.52\)) are discarded. Electrons are required to satisfy a loose likelihood identification requirement [72, 73] based on calorimeter shower shapes and track properties. The significance of the transverse impact parameter of the electron track is required to be less than five.

Muon candidates are reconstructed in the region \(|\eta | < 2.5\) from muon spectrometer tracks matching ID tracks. Muons are required to have \(p_{\text {T}} > 10 {\mathrm {\ GeV}}\) and pass the medium identification requirements defined in Ref. [74], based on the number of hits in the ID and muon spectrometer, and the compatibility of the charge-to-momentum ratio measured in the two detector systems.

Hadronically decaying tau leptons are reconstructed [75] from anti-\(k_t\) jets with \(E_\mathrm {T}\ge 10{\mathrm {\ GeV}}\) and \(|\eta |<2.5\) calibrated with a local cluster weighting technique [76]. Tau candidates are built from clusters of calorimeter cells within a cone of size \(\Delta R \equiv \sqrt{\smash [b]{(\Delta \eta )^2+(\Delta \phi )^2}}=0.2\) centred on the jet axis. Tau energy scale corrections are applied to subtract the energy originating from pile-up interactions and correct for the calorimeter response. Tau leptons are required to have \(p_{\text {T}} >20 {\mathrm {\ GeV}}\), and candidates reconstructed within the transition region \(1.37< |\eta | <1.52\) are discarded. Tau leptons are required to have either one or three associated tracks, with a charge sum of \(\pm 1\). A boosted-decision-tree discriminant is used to separate jets from tau leptons. It relies on track variables from the inner detector as well as longitudinal and transverse shower-shape variables from the calorimeters. The analysis makes use of loose and medium tau leptons, corresponding to identification working points with efficiencies of 60 and 55% for one-track tau leptons, respectively, and 50 and 40% for three-track tau leptons, respectively. Electrons mis-identified as 1-track tau leptons are rejected by imposing a \(p_{\text {T}}\)- and \(|\eta |\)-dependent requirement on the electron likelihood, which provides a constant efficiency of 95% for real tau leptons, with an inverse background efficiency (rejection factor) for electrons ranging from 30 to 150 depending on the \(|\eta |\) region.

The missing transverse momentum vector \(\vec {{p}}_\mathrm {T}^\mathrm {~miss}\), whose magnitude is denoted by \(E_{\mathrm {T}}^{\mathrm {miss}}\), is defined as the negative vector sum of the transverse momenta of all identified physics objects (electrons, muons, jets, and tau leptons) and an additional soft-term. Therefore, the \(E_{\mathrm {T}}^{\mathrm {miss}}\) calculation benefits from the dedicated calibration for each final-state object. The soft-term is constructed from all the tracks with \(p_{\text {T}} >400{\mathrm {\ MeV}}\) originating from the primary vertex which are not associated with any physics object. This track-based definition makes the soft-term largely insensitive to pile-up [77].

After object reconstruction, an overlap-removal procedure is applied to remove ambiguities in case the same object is reconstructed by several algorithms. Tau candidates are discarded if they are found to overlap with a light lepton (electron or muon). If an electron and a muon are reconstructed using the same inner-detector track, the electron is discarded. For overlapping jets and electrons, the electron is kept. Light leptons in the vicinity of a jet are considered to originate from secondary decays within the jet and are discarded. Finally, in case a jet is reconstructed as a tau lepton, the tau candidate is retained. The successive steps of this procedure are summarised in Table 1. The final-state objects considered in the analysis are those surviving the overlap-removal algorithm.

Table 1 Overview of the successive steps in the overlap removal algorithm

5 Event selection

The trigger used in this analysis is the lowest-threshold missing transverse momentum trigger that was active without restrictions during the whole 2015 data-taking period. The efficiency of that trigger is measured using data collected by a set of single-jet triggers, with events containing at least two jets and one loose tau lepton. The trigger is found to have an efficiency greater than 99% when requiring \(E_{\mathrm {T}}^{\mathrm {miss}} > 180 {\mathrm {\ GeV}}\) and a jet with \(p_{\text {T}} > 120 {\mathrm {\ GeV}}\) in the offline selection, which is referred to as trigger plateau conditions.

After trigger requirements, a pre-selection common to both channels is applied to ensure that only well-reconstructed events enter the analysis. Events containing no reconstructed primary vertex, mis-measured muon tracks, cosmic muon candidates, jets originating from calorimeter noise or reconstructed near inactive areas in the calorimeter, are vetoed. The presence of a second jet with \(p_{\text {T}} > 20 {\mathrm {\ GeV}}\) is required. To suppress the contribution from multi-jet events where large \(E_{\mathrm {T}}^{\mathrm {miss}}\) would arise from jet energy mis-measurement, a minimum angular separation in the transverse plane is imposed between either of the two leading jets and the missing transverse momentum, \(\Delta \phi (\text {jet}_{1,2},\vec {{p}}_\mathrm {T}^\mathrm {~miss})>0.4\).

As part of the pre-selection in the \(1\tau \) channel, events are required to contain exactly one medium tau lepton and no light lepton. The veto against electrons and muons was used in Run 1 to ensure the \(1\tau \) channel did not overlap with the \(e\tau \) and \(\mu \tau \) channels, which are not part of the present analysis. The lepton veto does not affect the expected sensitivity to the simplified model, and has been kept in the \(1\tau \) channel. Events with two or more loose tau leptons are rejected to make the \(1\tau \) and \(2\tau \) channels statistically independent. In the \(2\tau \) channel, at least two loose tau leptons are required at pre-selection level. No veto is applied against light leptons, as this would cause a sizeable selection inefficiency for GMSB signals.

In each channel, signal regions (SRs) are defined for several signal scenarios. The following kinematic variables are found to provide discrimination between signal and background, or between backgrounds themselves:

  • the transverse mass of the system formed by \(\vec {{p}}_\mathrm {T}^\mathrm {~miss}\) and a lepton \(\ell \) assumed to be massless,

    $$\begin{aligned} m^\ell _{\mathrm {T}} \equiv m_\mathrm {T}(\ell ,\vec {{p}}_\mathrm {T}^\mathrm {~miss})=\sqrt{2p_{\text {T}} ^{\ell }E_{\mathrm {T}}^{\mathrm {miss}} (1-\cos \Delta \phi (\ell ,\vec {{p}}_\mathrm {T}^\mathrm {~miss}))}~. \end{aligned}$$
    (2)

    For events where the missing transverse momentum and the lepton originate from a \(W\rightarrow \ell \nu \) decay, the \(m^\ell _{\mathrm {T}} \) distribution exhibits a Jacobian peak at the W boson mass. In the \(1\tau \) channel, the \(m_{\mathrm {T}} ^\tau \) variable is used. In the \(2\tau \) channel, the \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2}\) variable based on the two leading tau leptons is used, and \(m_\mathrm {T}^\mu \) is used for specific selections requiring a muon;

  • the scalar sum of the transverse momenta of all tau leptons and jets in the event, \(H_{\mathrm {T}} =\sum \limits _i p_{\text {T}} ^{\mathrm {\tau }_i}+ \sum \limits _j p_{\text {T}} ^{\mathrm {jet}_j}\) ;

  • the magnitude of the missing transverse momentum, \(E_{\mathrm {T}}^{\mathrm {miss}}\);

  • the effective mass, \(m_{\mathrm {eff}} =H_{\mathrm {T}} +E_{\mathrm {T}}^{\mathrm {miss}} \);

  • the \(m_\text {T2} ^{\tau \tau }\) variable [78, 79], also called stransverse mass, computed as

    $$\begin{aligned} m_\text {T2} ^{\tau \tau } =\sqrt{ \min \limits _{{\vec {p}_\mathrm {T}^{\,a}+ \vec {p}_\mathrm {T}^{\, b} = \vec {{p}}_\mathrm {T}^\mathrm {~miss}}}(\max [m_\mathrm {T}^2(\tau _1,\vec {p}_\mathrm {T}^{~a}) , m_\mathrm {T}^2(\tau _2,\vec {p}_\mathrm {T}^{~b})])}, \end{aligned}$$
    (3)

    where (ab) refers to two invisible particles that are assumed to be produced with transverse momentum vectors \(\vec {p}_\mathrm {T}^{~a,b} \). In this calculation, (ab) are assumed to be massless. The \(m_\text {T2} ^{\tau \tau }\) distribution has a kinematic endpoint for processes where massive particles are pair-produced, each particle decaying to a tau lepton and an undetected particle. In cases where multiple tau leptons are produced in a decay chain, there is no a-priori way to select the pair leading to the desired characteristic. For events with more than two tau-lepton candidates, \(m_\text {T2} ^{\tau \tau }\) is hence calculated using all possible tau-lepton pairs and the largest value is chosen;

  • the sum of the transverse masses of all jets and of the two leading tau leptons in the event, \(m_{\mathrm {T}} ^\mathrm {sum} = m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2} + \sum \limits _i m_{\mathrm T}^{\mathrm {jet}_i}\) , where \(m_{\mathrm T}^{\mathrm {jet}}\) is defined analogously to \(m^\ell _{\mathrm {T}} \) ;

  • the total number of jets, \(N_\text {jet}\);

  • the number of b-tagged jets, \(N_{b\text {-jet}}\).

These variables are also used for the selection of control regions (CRs) and validation regions (VRs) in the context of background modelling studies, as discussed in Sect. 6.

Figure 2 shows example kinematic distributions at pre-selection level. The dominant backgrounds are \(W(\tau \nu )\)+jets and \(t\bar{t}\) production in the \(1\tau \) channel, with subdominant contributions from \(Z(\nu \nu )\)+jets and \(Z(\tau \tau )\)+jets. In the \(2\tau \) channel, the pre-selection is dominated by \(t\bar{t}\), \(W(\tau \nu )\)+jets, and \(Z(\tau \tau )\)+jets events. The multi-jet and diboson backgrounds respectively contribute to about 0.2 and 2% of the total background in the \(1\tau \) channel, while their contribution in the \(2\tau \) channel amounts to 1 and 5%, respectively.

Fig. 2
figure 2

Kinematic distributions at pre-selection level, for the \(1\tau \) and \(2\tau \) channels. The last bin includes overflow events. The shaded bands indicate the statistical uncertainties in the background predictions. The contribution labelled as “other” includes diboson and multi-jet events, and the V+jets processes not explicitly listed in the legend. Red arrows in the data/SM ratio indicate bins where the corresponding entry falls outside the plotted range

In the \(1\tau \) channel, three characteristic regions in the \((m_{\tilde{g}},m_{\tilde{\chi }_1^0})\) parameter space of the simplified model are chosen as benchmark scenarios for the SR optimisation, with small (\(<100{\mathrm {\ GeV}}\)), medium (500–\(900{\mathrm {\ GeV}}\)) and large (\(>1200 {\mathrm {\ GeV}}\)) mass splittings between the gluino and the LSP. In the following, the associated SRs are called Compressed, Medium-Mass and High-Mass SRs, respectively. The Compressed SR exploits topologies where a high-\(p_{\text {T}}\) jet from initial-state radiation (ISR) recoils against the pair of gluinos. In this situation, the soft visible particles produced in the gluino decay receive a transverse Lorentz boost. While tau leptons and jets from gluino decays typically have low \(p_{\text {T}}\), such events have substantial \(E_{\mathrm {T}}^{\mathrm {miss}}\) since both LSPs tend to be emitted oppositely to the ISR jet in the transverse plane. In the case of large mass splitting, high-\(p_{\text {T}}\) jets come mainly from gluino decays, and a high-\(p_{\text {T}}\) requirement on the first two leading jets is effective in rejecting background without inducing a large inefficiency for the signal. A requirement on the transverse mass \(m_{\mathrm {T}} ^\tau \) is applied in the Medium-Mass and High-Mass SRs to suppress \(W(\tau \nu )\)+jets events as well as semileptonic \(t\bar{t}\) events with a tau lepton in the final state. The \(H_{\mathrm {T}}\) variable is also used in these two SRs, as \(H_{\mathrm {T}}\) increases for signal events with increasing mass splittings. No SR is defined for the GMSB model, as the expected sensitivity in the \(1\tau \) channel is significantly lower than that in the \(2\tau \) channel.

In the \(2\tau \) channel, two SRs are defined for the simplified model to cover small (\(<900{\mathrm {\ GeV}}\)) and large (\(>1200 {\mathrm {\ GeV}}\)) mass-splitting scenarios. The Compressed SR imposes a requirement on \(m_\text {T2} ^{\tau \tau }\) to exploit the fact that most of SM background contributions exhibit a kinematic endpoint around the W or Z boson mass, which is not the case for tau leptons produced in the cascade decay of gluinos. A requirement is also applied on \(m_{\mathrm {T}} ^\mathrm {sum}\) to take advantage of the large \(E_{\mathrm {T}}^{\mathrm {miss}}\) and the large jet and tau lepton multiplicity that is expected for signal events. The High-Mass SR includes a requirement on \(H_{\mathrm {T}}\), which is efficient for High-Mass gluino signals. A requirement on \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2}\) is also applied. In \(Z(\tau \tau )\)+jets events, where \(E_{\mathrm {T}}^{\mathrm {miss}}\) originates from neutrinos from tau lepton decays, the trigger plateau requirement selects events with a high-\(p_{\text {T}}\) Z boson recoiling against jets in the transverse plane. This topology leads to tau leptons with a small \(\Delta \phi \) separation, which results in low values of \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2}\) given that tau neutrinos are themselves collimated with the visible decay products of tau leptons. For dileptonic \(t\bar{t}\) events with two tau leptons and large genuine \(E_{\mathrm {T}}^{\mathrm {miss}}\), and for \(W(\tau \nu )\)+jets events and semileptonic \(t\bar{t}\) events with a high-\(p_{\text {T}}\) jet mis-identified as a tau lepton, \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2}\) can be larger, but even larger values are expected for High-Mass gluino signals. A signal region is also defined for the GMSB model, and targets more specifically squark–antisquark production rather than gluino pair production in the region \(\Lambda \gtrsim 80 {\mathrm {\ TeV}}\), not excluded by Run 1 searches. Among the distinctive features that give large discrimination power to \(H_{\mathrm {T}}\), decay chains are potentially longer than in the simplified model, and the almost-massless gravitino LSP leaves more phase space to other particles in the decay. Table 2 summarises the selection criteria for all the SRs of the \(1\tau \) and \(2\tau \) channels.

Table 2 Selection criteria for the signal regions (SRs) of the \(1\tau \) channel (top) and the \(2\tau \) channel (bottom)

6 Background estimation

To predict the background contributions in the SRs, the normalisation of the dominant backgrounds is fitted to data in dedicated CRs. In each channel and for each SR, a simultaneous fit over the relevant CRs is performed using HistFitter [80] to extract these normalisation factors. Control regions are designed to have an enhanced contribution from a single background process, with contributions from other backgrounds as small as possible to reduce the uncertainties originating from correlations between CRs. Furthermore, CRs are defined in phase-space regions close to that of SRs, to avoid the extrapolation of the background normalisations over very different kinematic regimes.

A set of VRs is defined in intermediate phase-space regions between a SR and its associated CRs, where signal contributions are small. These VRs are not part of the fit; they are used to compare the fitted background predictions with the observed data in the vicinity of the SRs before unblinding those.

6.1 Vector-boson and top quark backgrounds

In both channels, the dominant backgrounds originate from SM processes involving the top quark or a massive vector boson and jets. These two backgrounds can be separated from each other by either requiring or vetoing the presence of a b-tagged jet in the event. In addition, a tau-lepton candidate can be either a genuine tau lepton (true tau lepton) from a \(W(\tau \nu )\) decay or a jet mis-identified as a tau lepton (fake tau lepton), which leads to two types of CRs.

In CRs targeting true tau-lepton contributions, the normalisation factor is used to absorb the theoretical uncertainties in cross-section computations, the experimental uncertainties in the integrated luminosity, and potential differences in the tau-lepton reconstruction and identification efficiencies between data and simulation. In the case of fake tau-lepton contributions, the normalisation factor combines several effects: the quark/gluon composition of jets mis-identified as tau leptons, which is process-dependent, the parton shower and hadronisation models of the generator, and the modelling in the simulation of jet shower shapes in the calorimeter, which mainly depends on the Geant4 hadronic interaction model and the modelling of the ATLAS detector. Other contributions affecting the background normalisation include the modelling of the kinematics and acceptance of background processes. These contributions are absorbed into the true and fake tau-lepton normalisation factors in the \(1\tau \) channel, while they are treated as a separate kinematic normalisation factor in the \(2\tau \) channel to avoid double-counting (true- or fake-tau normalisation factors are applied to each tau lepton, while the kinematic normalisation factor is applied once per event).

In the \(1\tau \) channel, four CRs are defined, with four associated normalisation factors. They target the top quark background (including \(t\bar{t}\) and single-top-quark processes) with either a true or a fake tau lepton, and V+jets events with either a true or a fake tau lepton, respectively dominated by \(W(\tau \nu )\)+jets and \(Z(\nu \nu )\)+jets processes. The discrimination between true and fake tau-lepton contributions is achieved by a requirement on \(m_\mathrm {T}^\tau \). A common set of four CRs is defined for the Medium-Mass and High-Mass SRs, due to the similarity of background compositions and event kinematics. These are separated from the SRs by an upper bound on \(m_{\mathrm {T}} ^\tau \). Another set of four CRs is defined for the Compressed SR, to study more specifically the background modelling for low-\(p_{\text {T}}\) tau leptons. These CRs are separated from the SR by an upper bound on \(E_{\mathrm {T}}^{\mathrm {miss}}\). The selection criteria defining these CRs are summarised in Tables 3 and 4. Figure 3 illustrates the good modelling of the background in the various CRs after the fit.

Three types of VRs are used in the \(1\tau \) channel for the Medium- and High-Mass SRs, to validate the background extrapolation from low-\(H_{\mathrm {T}}\) to high-\(H_{\mathrm {T}}\) for selections based on true tau leptons, and the extrapolations along \(H_{\mathrm {T}}\) and \(m_{\mathrm {T}} ^\tau \) for selections based on fake tau leptons. The separation of the VRs from the SRs is achieved by inverting the selections on \(m_{\mathrm {T}} ^\tau \) or \(H_{\mathrm {T}}\). For the Compressed SR, four VRs are used to validate the extrapolation of V+jets and top quark background predictions along \(E_{\mathrm {T}}^{\mathrm {miss}}\), for both the true and fake tau-lepton selections. The separation of the VRs from the SRs is achieved by an inverted requirement on \(m_{\mathrm {T}} ^\tau \) and \(E_{\mathrm {T}}^{\mathrm {miss}}\), for true and fake tau-lepton VRs, respectively. The selection criteria defining all the VRs in the \(1\tau \) channel are listed in Tables 3 and 4.

Table 3 Selection criteria defining the control and validation regions of the Compressed SR in the \(1\tau \) channel. The pre-selection criteria are also applied although not shown in the table. Unless mentioned, criteria apply to both the top quark and V+jets background regions. Symbols* indicate that criteria are only applied to regions targeting the top quark background, and symbols\(^{\dag }\) denote criteria only applied to V+jets background regions
Table 4 Selection criteria defining the control and validation regions of the Medium-Mass and High-Mass SRs in the \(1\tau \) channel. The pre-selection criteria are also applied although not shown in the table. Unless mentioned, criteria apply to both the top quark and V+jets background regions. Symbols* indicate that criteria are only applied to regions targeting the top quark background, and symbols\(^{\dag }\) denote criteria only applied to V+jets background regions
Fig. 3
figure 3

Missing transverse momentum and leading-jet \(p_{\text {T}}\) distributions in two different control regions of the \(1\tau \) channel after the fit, illustrating the overall background modelling in the CRs. By construction, the total fitted background is equal to the number of observed events in each CR. The last bin includes overflow events. The shaded bands indicate the statistical uncertainties in the background predictions. Red arrows in the data/SM ratio indicate bins where the corresponding entry falls outside the plotted range

In the \(2\tau \) channel, the dominant backgrounds are \(Z(\tau \tau )\)+jets and dileptonic top quark contributions (including \(t\bar{t} \) and single-top quark processes) with two true tau leptons, \(W(\tau \nu )\)+jets and semileptonic top quark contributions with one true and one fake tau lepton, and \(W(\ell \nu )\)+jets and top quark contributions with two fake tau leptons. Control regions are defined for W+jets and top quark backgrounds to extract normalisation factors related to the modelling of the process kinematics, and the modelling of real and fake tau leptons in the simulation. These CRs are separated from the SRs by replacing the requirement on the two tau leptons with requirements on different final-state objects, which stand in for true or fake tau leptons. To be independent from tau-lepton considerations, kinematic CRs are based on events with one muon, jets, \(E_{\mathrm {T}}^{\mathrm {miss}}\), and without or with b-jets, to select \(W(\mu \nu )\)+jets and semileptonic top quark events with a final-state muon, respectively. The fake tau-lepton CRs use the same baseline selections as the kinematic CRs, but in addition, the presence of a loose tau-lepton candidate is required. Events with large \(m_\mathrm {T}^\mu \) values are discarded to suppress the dileptonic top quark background with a muon and a true tau lepton. The true tau-lepton CRs, which target \(W(\tau \nu )\)+jets and semileptonic top quark processes with a true tau lepton, are based on events with a loose tau lepton, jets and \(E_{\mathrm {T}}^{\mathrm {miss}}\), without or with b-jets. Similarly, contributions from fake tau leptons are suppressed by a requirement on \(m_\mathrm {T}^\tau \). A separate CR is designed to study \(Z(\tau \tau )\)+jets events by inverting the \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2}\) and \(H_{\mathrm {T}}\) requirements from the SRs. This selection requires two loose tau leptons of opposite electric charge. The selection criteria defining the various CRs are summarised in Table 5. Figure 4 illustrates the good background modelling in the CRs after the fit.

The VRs of the \(2\tau \) channel are presented in Table 6. For the \(Z(\tau \tau )\)+jets background, the validation of the background extrapolation is performed from low-\(H_{\mathrm {T}}\) to high-\(H_{\mathrm {T}}\), while keeping the upper bound on \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2} \) which is effective in selecting \(Z(\tau \tau )\)+jets events. The validity of the top quark and \(W(\tau \nu )\)+jets background predictions obtained with alternative object selections are checked for selections with two reconstructed tau leptons. High values of \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2} \) are required to suppress \(Z(\tau \tau )\)+jets events as in the SRs, while upper bounds on \(m_\text {T2}\) and \(H_{\mathrm {T}}\) ensure there is no overlap between SRs and VRs. The same set of CRs and VRs is used for the three SRs of the \(2\tau \) channel.

The resulting normalisation factors for both channels do not deviate from 1 by more than 25%, except for \(t\bar{t}\) events with fake tau leptons in the \(1\tau \) channel, where the normalisation factor reaches 2 within large statistical uncertainty. The typical level of agreement between data and background distributions in the CRs after the fit can be seen in Figs. 3 and 4. Good modelling of kinematic distributions is observed in all CRs. The comparison between the number of observed events and the predicted background yields in the VRs is displayed in Fig. 5. Agreement within approximately one standard deviation is observed.

Table 5 Overview of the control regions used in the \(2\tau \) channel
Fig. 4
figure 4

\(H_{\mathrm {T}}\) distribution in the top quark fake-tau CR and transverse momentum of the leading tau lepton in the \(Z(\tau \tau )\)+jets CR of the \(2\tau \) channel after the fit, illustrating the overall background modelling in the control regions. By construction, the total fitted background is equal to the number of observed events in each CR. The last bin includes overflow events. The shaded bands indicate the statistical uncertainties in the background predictions

Table 6 Overview of the validation regions defined in the \(2\tau \) channel
Fig. 5
figure 5

Number of observed events, \(n_\mathrm {obs}\), and predicted background yields after the fit, \(n_\mathrm {pred}\), in the validation regions of the \(1\tau \) and \(2\tau \) channels. The background predictions are scaled using normalisation factors derived in the control regions. The shaded bands indicate the statistical uncertainties in the background predictions, and correspond to the \(\sigma _\mathrm {tot}\) uncertainties used in the lower part of the figure

6.2 Multi-jet background

The multi-jet background contributes to the selection when two conditions are simultaneously fulfilled: jets have to be mis-identified as tau leptons, and large missing transverse momentum must arise from jet energy mis-measurement. This background is estimated from data, because final-state objects arising from mis-measurements are much more challenging to simulate than the reconstruction and identification of genuine objects. Moreover, the very large multi-jet production cross section at the LHC would imply simulating a prohibitively large number of multi-jet events.

The jet smearing method [81] employed in the \(1\tau \) channel proceeds in two steps. First, multi-jet events with well-measured jets are selected in a data sample collected by single-jet triggers. This is achieved by requiring \(E_{\mathrm {T}}^{\mathrm {miss}}/(\sum E_\mathrm {T})^{1/3} < 5 {\mathrm {\ GeV}}^{2/3}\), where the objects entering the \(\sum E_\mathrm {T}\) term are those entering the \(E_{\mathrm {T}}^{\mathrm {miss}}\) calculation. Selected events are required to have at least two jets, no light lepton and exactly one tau candidate satisfying the medium identification criteria. The selection is dominated by multi-jet production, such that most tau candidates are jets mis-identified as tau leptons. In a second step, jet energies are smeared according to the \(p_{\text {T}}\)-dependent jet energy resolution extracted from simulation. The smearing is performed multiple times for each event, leading to a large pseudo-data set where \(E_{\mathrm {T}}^{\mathrm {miss}}\) originates from resolution effects and which includes an adequate fraction of jets mis-identified as tau leptons.

This method cannot be used in the \(2\tau \) channel because of the limited number of events with well-measured jets that contain at least two loose tau candidates. Instead, a fake rate approach is adopted. The probability for jets to be mis-identified as tau leptons or muons, obtained from simulated dijet events, is applied to jets from an inclusive data sample collected by single-jet triggers and dominated by multi-jet events.

7 Systematic uncertainties

For all simulated processes, theoretical and experimental systematic uncertainties are considered. The former includes cross-section uncertainties, which are not relevant for the dominant backgrounds normalised to data, and generator modelling uncertainties. The latter refers to all the uncertainties related to the reconstruction, identification, calibration and corrections applied to jets, tau leptons, electrons, muons and missing transverse momentum. Specific uncertainties are evaluated for the multi-jet background, which is estimated from data.

Theoretical uncertainties are evaluated for all simulated samples. For backgrounds that are normalised in CRs, the uncertainty in the transfer factors, i.e. the ratio of the expected event yields in a SR or VR over the respective CR, is evaluated for all SRs and VRs. The difference between the nominal simulation and the systematically varied sample is used as an additional uncertainty. For backgrounds that are evaluated from simulation alone, i.e. the diboson background, a global normalisation uncertainty is added. Uncertainties for V+jets samples generated with Sherpa are estimated by up and down variations by factors of two and one-half in the renormalisation and factorisation scales, resummation scale (maximum scale of the additional emission to be resummed by the parton shower) and CKKW matching scale (matching between matrix elements and parton shower). The effect of scale variations is parameterised at generator level as a function of the vector boson \(p_{\text {T}}\) and the number of particle jets. For the top quark background, the nominal predictions from Powheg-Box + Pythia6 are compared with predictions from alternative generators, and the differences are taken as systematic uncertainties. The MG5_aMC@NLO + Herwig++ generators are used to evaluate uncertainties in the modelling of the hard scattering, parton shower and hadronisation. The Powheg-Box + Herwig++ generators are used to compute a specific uncertainty in the parton shower and hadronisation models. An uncertainty in the ISR modelling is also assessed by varying the Powheg-Box parameter which controls the transverse momentum of the first additional parton emission beyond the Born configuration. In the case of the diboson background, a 6% uncertainty in the cross section due to scale and PDF uncertainties is considered. Uncertainties in signal cross sections are obtained by using different PDF sets and factorisation and renormalisation scales, as discussed in Section 3.

Systematic uncertainties affecting jets arise from the jet energy scale and jet energy resolution [66], as well as efficiency corrections for jet-vertex-tagging [68] and b-tagging [82]. A set of \(p_{\text {T}}\)- and \(\eta \)-dependent uncertainties in the jet energy scale and resolution is estimated by varying the conditions used in the simulation. Another set of uncertainties accounts for the modelling of the residual pile-up dependence. Additional uncertainties account for the jet flavour composition of samples that are used to derive in-situ energy scale corrections, where jets are calibrated against well-measured objects. A punch-through uncertainty for jets not entirely contained in the calorimeters, as well as a single-hadron response uncertainty, are also included for high-\(p_{\text {T}}\) jets. An overall uncertainty in the jet energy resolution is applied to jets in the simulation as a Gaussian energy smearing.

Systematic uncertainties affecting correctly identified tau leptons arise from the reconstruction, identification and tau-electron overlap-removal efficiencies, and the energy scale calibration [75]. Most of the uncertainties are estimated by varying nominal parameters in the simulation: detector material, underlying event, hadronic shower model, pile-up and noise in the calorimeters. The uncertainty in the energy scale also includes non-closure of the calibration found in simulation, a single-pion response uncertainty, and an uncertainty in the in-situ energy calibration of data with respect to simulation derived in \(Z(\tau \tau )\) events with a hadronically decaying tau lepton and a muon in the final state. In the case of signal samples, which undergo fast calorimeter simulation, a dedicated uncertainty takes into account the difference in performance between full and fast simulation. The effect of mis-identified tau leptons is largely constrained by the background estimation approaches. Uncertainties arise due to the extrapolation from the CRs to the VRs and SRs. These are considered as part of the theory uncertainties, which account for the impact of hadronisation on the mis-identification of jets as tau leptons.

Systematic uncertainties affecting electrons and muons are related to the energy or momentum calibration, as well as efficiency corrections for the reconstruction, identification and isolation requirements. These uncertainties have a negligible impact on the background predictions.

Systematic uncertainties in the missing transverse momentum originate from uncertainties in the energy or momentum calibration of jets, tau leptons, electrons, and muons, which are propagated to the \(E_{\mathrm {T}}^{\mathrm {miss}}\) calculation. Additional uncertainties are related to the calculation of the track-based soft-term. These uncertainties are derived by studying the \(p_{\text {T}}\) balance in \(Z(\mu \mu )\) events between the soft-term and the hard-term composed of all reconstructed objects. Soft-term uncertainties include scale uncertainties along the hard-term axis, and resolution uncertainties along and perpendicular to the hard-term axis [83].

A systematic uncertainty of the pile-up modelling is estimated by varying the distribution of the average number of interactions per bunch crossing in the simulation. The range of the variation is determined by studying the correlation in data and simulation between the average number of interactions and the number of reconstructed primary vertices. This uncertainty ranges from a few percent in the \(1\tau \) channel to about 15% in the poorly populated SRs of the \(2\tau \) channel.

The uncertainty in the integrated luminosity is \(\pm 2.1\)%. It is derived, following a methodology similar to that detailed in Ref. [84], from a calibration of the luminosity scale using xy beam-separation scans performed in August 2015. This uncertainty only affects the diboson background prediction and the signal yields, as other backgrounds are normalised to the data.

The systematic uncertainty in the small multi-jet background contribution is estimated by considering alternative normalisation regions and different jet smearing parameters in the case of the \(1\tau \) channel. An uncertainty of the order of 70% is found for the \(1\tau \) channel, and a 100% uncertainty is assigned in the \(2\tau \) channel.

The influence of the main systematic uncertainties in the total background predictions in the SRs of the \(1\tau \) and \(2\tau \)   channels are summarised in Table 7. The uncertainties reported in the table are derived assuming that no signal is present in the CRs.

The total uncertainties range between 13 and 72%. For all SRs, the largest uncertainties, between 7 and 60%, originate from the MC generator modelling of top quark events. The larger uncertainty found for the Compressed SR of the \(2\tau \) channel is explained by the larger statistical uncertainty in the predictions from alternative generators. Energy scale uncertainties affecting tau leptons and jets contribute significantly in all regions as well. Other uncertainties, e.g. in the b-tagging efficiency and the jet energy resolution, do not play a large role in most of the SRs.

Table 7 Dominant systematic uncertainties in the total background predictions, in percent, for the signal regions of the \(1\tau \) (top) and \(2\tau \) (bottom) channels. The total systematic uncertainty accounts for other minor contributions not listed in this table, as well as correlations between the uncertainties

8 Results

Kinematic distributions for extended SR selections of the \(1\tau \) and \(2\tau \) channels are shown in Figs. 6 and  7, respectively. These regions are defined by the same set of selection criteria as for the SRs, except that the criterion corresponding to the plotted variable is not applied. Data and background predictions, fitted to data in the control regions, are compared, and signal predictions are also shown for several benchmark models. Variables providing the most discrimination between signal and background are displayed: \(m_{\mathrm {T}} ^\tau \) distributions for the \(1\tau \) channel, and \(m_{\mathrm {T}} ^\mathrm {sum}\), \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2}\) and \(H_{\mathrm {T}}\) distributions for the \(2\tau \) channel. Reasonable agreement between data and background distributions is observed, given the low event yields remaining in data after these selections. The numbers of observed events and expected background events in all SRs are reported in Tables 8 and 9 for the \(1\tau \) and \(2\tau \) channels, respectively.

Fig. 6
figure 6

\(m_\mathrm {T}^\tau \) distributions for “extended SR selections” of the \(1\tau \) channel, for a the Compressed SR selection without the \(m_{\mathrm {T}} ^\tau > 80{\mathrm {\ GeV}}\) requirement, b the Medium-Mass SR selection without the \(m_{\mathrm {T}} ^\tau > 200{\mathrm {\ GeV}}\) requirement, and c the High-Mass SR selection without the \(m_{\mathrm {T}} ^\tau >200 {\mathrm {\ GeV}}\) requirement. The last bin includes overflow events. The shaded bands indicate the statistical uncertainties in the background predictions. Red arrows in the data/SM ratio indicate bins where the corresponding entry falls outside the plotted range. The signal region is indicated by the black arrow. Signal predictions are overlaid for several benchmark models, normalised to their predicted cross sections. For the simplified model, LM low mass splitting, or compressed scenario, with \(m_{\tilde{g}}=665 {\mathrm {\ GeV}}\) and \(m_{\tilde{\chi }_1^0}=585 {\mathrm {\ GeV}}\); MM medium mass splitting, with \(m_{\tilde{g}}=1145 {\mathrm {\ GeV}}\) and \(m_{\tilde{\chi }_1^0}=265 {\mathrm {\ GeV}}\); HM high mass splitting scenario, with \(m_{\tilde{g}}=1305 {\mathrm {\ GeV}}\) and \(m_{\tilde{\chi }_1^0}=105 {\mathrm {\ GeV}}\)

Fig. 7
figure 7

Kinematic distributions for “extended SR selections” of the \(2\tau \) channel, for a \(m_{\mathrm {T}} ^\mathrm {sum}\) in the Compressed SR selection without the \(m_{\mathrm {T}} ^\mathrm {sum}>1400 {\mathrm {\ GeV}}\) requirement, b \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2}\) in the High-Mass SR selection without the \(m_{\mathrm T}^{\tau _1}+m_{\mathrm T}^{\tau _2} >350{\mathrm {\ GeV}}\) requirement, and c \(H_{\mathrm {T}}\) in the GMSB SR selection without the \(H_{\mathrm {T}} > 1700 {\mathrm {\ GeV}}\) requirement. The last bin includes overflow events. The shaded bands indicate the statistical uncertainties in the background predictions. The signal region is indicated by the black arrow. Signal predictions are overlaid for several benchmark models, normalised to their predicted cross sections. For the simplified model, MM medium mass splitting, with \(m_{\tilde{g}}=1145 {\mathrm {\ GeV}}\) and \(m_{\tilde{\chi }_1^0}=265 {\mathrm {\ GeV}}\); HM high mass splitting scenario, with \(m_{\tilde{g}}=1305 {\mathrm {\ GeV}}\) and \(m_{\tilde{\chi }_1^0}=105 {\mathrm {\ GeV}}\). The GMSB benchmark model corresponds to \(\Lambda = 90{\mathrm {\ TeV}}\) and \(\tan \beta =40\)

No significant excess is observed in data over the SM predictions. Therefore, upper limits are set at the 95% confidence level (CL) on the number of hypothetical signal events, or equivalently, on the signal cross section. The one-sided profile-likelihood-ratio test statistic is used to assess the compatibility of the observed data with the background-only and signal-plus-background hypotheses. Systematic uncertainties are included in the likelihood function as nuisance parameters with Gaussian probability densities. Following the standards used for LHC analyses, p-values are computed according to the \(\mathrm {CL_s}\) prescription [85].

Model-independent upper limits are calculated for each SR, assuming no signal contribution in the CRs. The results are derived using profile-likelihood-ratio distributions obtained from pseudo-experiments. Upper limits on signal yields are converted into limits on the visible cross section (\(\sigma _\mathrm {vis}\)) of BSM processes by dividing by the integrated luminosity of the data. The visible cross section is defined as the product of production cross section, acceptance and selection efficiency. Results are summarised at the bottom of Tables 8 and 9. The most stringent observed upper limits on the visible cross section are \(1.17~\mathrm {fb}\) for the High-Mass SR of the \(1\tau \) channel and \(1.07~\mathrm {fb}\) for the High-Mass and GMSB SRs of the \(2\tau \) channel.

Table 8 Number of observed events and predicted background yields in the three signal regions of the \(1\tau \)  channel. The background prediction is scaled using normalisation factors derived in the control regions. All systematic and statistical uncertainties are included in the quoted uncertainties. The bottom part of the table shows the observed and expected model-independent upper limits at the 95% CL on the number of signal events, \(S_\mathrm{obs}^{95}\) and \( S_\mathrm{exp}^{95} \), respectively, the corresponding observed upper limit on the visible cross section, \(\langle \mathrm{\sigma _{vis}}\rangle _\mathrm{obs}^{95}\), and the \(\mathrm {CL_b}\) value, i.e. the confidence level observed for the background-only hypothesis
Table 9 Number of observed events and predicted background yields in the three signal regions of the \(2\tau \) channel. The background prediction is scaled using normalisation factors derived in the control regions. All systematic and statistical uncertainties are included in the quoted uncertainties. The bottom part of the table shows the observed and expected model-independent upper limits at the 95% CL on the number of signal events, \(S_\mathrm{obs}^{95}\) and \( S_\mathrm{exp}^{95} \), respectively, the corresponding observed upper limit on the visible cross section, \(\langle \mathrm{\sigma _{vis}}\rangle _\mathrm{obs}^{95}\), and the \(\mathrm {CL_b}\) value, i.e. the confidence level observed for the background-only hypothesis

The results are interpreted in the context of the simplified model of gluino pair production and the GMSB model. In the case of model-dependent interpretations, the signal contribution in the control regions is included in the calculation of upper limits, and asymptotic properties of test-statistic distributions are used [86]. Exclusion contours at the 95% CL are derived in the \((m_{\tilde{g}},m_{\tilde{\chi }_1^0})\) plane for the simplified model and in the \((\Lambda ,\tan \beta )\) plane for the GMSB model. Results are shown in Figs. 8 and 9. The solid lines and the dashed lines correspond to the observed and median expected limits, respectively. The band shows the one-standard-deviation spread of the expected limits around the median, which originates from statistical and systematic uncertainties in the background and signal. The theoretical uncertainty in the signal cross section is not included in the band. Its effect on the observed limits is shown separately as the dotted lines. For the simplified model, exclusion contours are shown for the \(1\tau \) and \(2\tau \) channels and their combination. The combination is performed by selecting, for each signal scenario, the SR with the lowest expected \(\mathrm {CL_s}\) value. The combination retains the Compressed SR of the \(1\tau \) channel in the region where the LSP mass is close to the gluino mass, and favours the High-Mass SR of the \(2\tau \) channel when the mass splitting is large. For the GMSB model, limits are shown for the GMSB SR of the \(2\tau \) channel. The stronger limits at high values of \(\tan \beta \) are explained by the nature of the NSLP, which is the lightest scalar tau in this region. For both models, the exclusion limits obtained with 3.2 \({\mathrm{{fb}^{-1}}}\) of collision data at \(\sqrt{s}=13 {\mathrm {\ TeV}}\) significantly improve upon the previous ATLAS results [14, 19] established with 20.3 \({\mathrm{{fb}^{-1}}}\) of 8 \({\mathrm {\ TeV}}\) data.

Fig. 8
figure 8

Exclusion contours at the 95% confidence level for the simplified model of gluino pair production, based on results from the \(1\tau \) and \(2\tau \) channels. The red solid line and the blue dashed line correspond to the observed and median expected limits, respectively, for the combination of the two channels. The yellow band shows the one-standard-deviation spread of expected limits around the median. The effect of the signal cross-section uncertainty in the observed limits is shown as red dotted lines. Additionally, expected limits are shown for the \(1\tau \) and \(2\tau \) channels individually as dashed green and magenta lines, respectively. The previous ATLAS result [19] obtained with 20.3 \({\mathrm{{fb}^{-1}}}\) of \(8 {\mathrm {\ TeV}}\) data is shown as the grey filled area

Fig. 9
figure 9

Exclusion contours at the 95% confidence level for the gauge-mediated supersymmetry-breaking model, based on results from the \(2\tau \) channel. The red solid line and the blue dashed line correspond to the observed and median expected limits, respectively. The yellow band shows the one-standard-deviation spread of expected limits around the median. The effect of the signal cross-section uncertainty in the observed limits is shown as red dotted lines. The previous ATLAS result [14] obtained with 20.3 \({\mathrm{{fb}^{-1}}}\) of \(8 {\mathrm {\ TeV}}\) data is shown as the grey filled area

9 Summary

A search for squarks and gluinos has been performed in events with hadronically decaying tau leptons, jets and missing transverse momentum, using 3.2 \({\mathrm{{fb}^{-1}}}\) of pp collision data at \(\sqrt{s}=13\) TeV recorded by the ATLAS detector at the LHC in 2015. Two channels, with either one tau lepton or at least two tau leptons, are separately optimised. The numbers of observed events in the different signal regions are in agreement with the Standard Model predictions. Results are interpreted in the context of a gauge-mediated supersymmetry breaking model and a simplified model of gluino pair production with tau-rich cascade decay. In the GMSB model, limits are set on the SUSY-breaking scale \(\Lambda \) as a function of \(\tan \beta \). Values of \(\Lambda \) below \(92 {\mathrm {\ TeV}}\) are excluded at the 95% CL, corresponding to gluino masses below \(2000 {\mathrm {\ GeV}}\). A stronger exclusion is achieved for large values of \(\tan \beta \), where \(\Lambda \) and gluino mass values are excluded up to \(107 {\mathrm {\ TeV}}\) and \(2300 {\mathrm {\ GeV}}\), respectively. In the simplified model, gluino masses are excluded up to \(1570 {\mathrm {\ GeV}}\) for neutralino masses around \(100 {\mathrm {\ GeV}}\), neutralino masses below \(700 {\mathrm {\ GeV}}\) are excluded for all gluino masses between 800 and \(1500 {\mathrm {\ GeV}}\), while the strongest neutralino-mass exclusion of \(750 {\mathrm {\ GeV}}\) is achieved for gluino masses around \(1450 {\mathrm {\ GeV}}\). A dedicated signal region provides good sensitivity to scenarios with a small mass difference between the gluino and the neutralino LSP.