1 Introduction

With a mass of 1.777 \(\,\hbox {GeV}\) and a proper decay length of 87 \(\upmu \)m [1], tau leptons decay either leptonically (\(\tau \rightarrow \ell \nu _\ell \nu _\tau \), \(\ell =e, \mu \)) or hadronically (\(\tau \rightarrow \mathrm {hadrons} \; \nu _\tau \), denoted \(\tau _{\mathrm {had}}\)) and do so typically before reaching active regions of the ATLAS detector. They can thus only be identified via their decay products. In this paper, only hadronic tau lepton decays are considered. The hadronic tau lepton decays represent 65 % of all possible decay modes [1]. In these, the hadronic decay products are one or three charged pions in 72 and 22 % of all cases, respectively. Charged kaons are present in the majority of the remaining hadronic decays. In 78 % of all hadronic decays, up to one associated neutral pion is also produced. The neutral and charged hadrons stemming from the tau lepton decay make up the visible decay products of the tau lepton, and are in the following referred to as \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\).

The main background to hadronic tau lepton decays is from jets of energetic hadrons produced via the fragmentation of quarks and gluons. This background is already present at trigger level (also referred to as online in the following). Other important backgrounds are electrons and, to a lesser degree, muons, which can mimic the signature of tau lepton decays with one charged hadron. In the context of both the trigger and the offline event reconstruction (shortened to simply offline in the following), discriminating variables based on the narrow shower shape, the distinct number of charged particle tracks and the displaced tau lepton decay vertex are used.

Final states with hadronically decaying tau leptons are an important part of the ATLAS physics program. Examples are measurements of Standard Model processes [26], Higgs boson searches [7], searches for new physics such as Higgs bosons in models with extended Higgs sectors [810], supersymmetry (SUSY) [1113], heavy gauge bosons [14] and leptoquarks [15]. This places strong requirements on the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)identification algorithms (in the following, referred to as tau identification): robustness and high performance over at least two orders of magnitude in transverse momentum with respect to the beam axis (\(p_{\mathrm {T}}\)) of \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\), from about 15 \(\,\hbox {GeV}\) (decays of W and Z bosons or scalar tau leptons) to a few hundred \(\,\hbox {GeV}\) (SUSY Higgs boson searches) and up to beyond 1 \(\,\hbox {TeV}\) (\(Z'\) searches). At the same time, an excellent energy resolution and small energy scale uncertainty are particularly important where resonances decaying to tau leptons need to be separated (e.g. \(Z \rightarrow \tau \tau \) from \(H \rightarrow \tau \tau \) mass peaks). The triggering for final states which rely exclusively on tau triggers is particularly challenging, e.g. \(H \rightarrow \tau \tau \) where both tau leptons decay hadronically. At the trigger level, in addition to the challenges of offline tau identification, bandwidth and time constraints need to be satisfied and the trigger identification is based on an incomplete reconstruction of the event. The ATLAS trigger system, together with the detector and the simulation samples used for the studies presented, are briefly described in Sect. 2.

The ATLAS offline tau identification uses various discriminating variables combined in Boosted Decision Trees (BDT) [16, 17] to reject jets and electrons. The offline tau energy scale is set by first applying a local hadronic calibration (LC) [18] appropriate for a wide range of objects and then an additional tau-specific correction based on simulation. The online tau identification is implemented in three different steps, as is required by the ATLAS trigger system architecture [19]. The same identification and energy calibration procedures as for offline are used in the third level of the trigger, while the first and second trigger levels rely on coarser identification and energy calibration procedures. A description of the trigger and offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)reconstruction and identification algorithms is presented in Sect. 3, and the trigger and offline energy calibration algorithms are discussed in Sect. 5.

The efficiency of the identification and the energy scale are measured in dedicated studies using a \(Z \rightarrow \tau \tau \)-enhanced event sample of collision data recorded by the ATLAS detector [20] at the LHC [21] in 2012 at a centre-of-mass energy of 8 \(\,\hbox {TeV}\). This is described in Sects. 4 and 5. Conclusions and outlook are presented in Sect. 6.

Table 1 Tau triggers with their corresponding kinematic requirements. Examples of physics processes targeted by each trigger are also listed, where \(\tau _{\mathrm {had}}\) and \(\tau _{\mathrm {lep}}\) refer to hadronically and leptonically decaying tau leptons, respectively

2 ATLAS detector and simulation

2.1 The ATLAS detector

The ATLAS detector [20] consists of an inner tracking system surrounded by a superconducting solenoid, electromagnetic (EM) and hadronic (HAD) calorimeters, and a muon spectrometer (MS). The inner detector (ID) is immersed in a 2 T axial magnetic field, and consists of pixel and silicon microstrip (SCT) detectors inside a transition radiation tracker (TRT), providing charged-particle tracking in the region \(|\eta |<2.5\).Footnote 1 The EM calorimeter uses lead and liquid argon (LAr) as absorber and active material, respectively. In the central rapidity region, the EM calorimeter is divided in three layers, one of them segmented in thin \(\eta \) strips for optimal \(\gamma / \pi ^0\) separation, and completed by a presampler layer for \(|\eta | < 1.8\). Hadron calorimetry is based on different detector technologies, with scintillator tiles (\(|\eta | < 1.7\)) or LAr (\(1.5 < |\eta | < 4.9\)) as active medium, and with steel, copper, or tungsten as the absorber material. The calorimeters provide coverage within \(|\eta |<4.9\). The MS consists of superconducting air-core toroids, a system of trigger chambers covering the range \(|\eta |<2.4\), and high-precision tracking chambers allowing muon momentum measurements within \(|\eta |<2.7\).

Physics objects are identified using their specific detector signatures; electrons are reconstructed by matching a track from the ID to an energy deposit in the calorimeters [22, 23], while muons are reconstructed using tracks from the MS and ID [24]. Jets are reconstructed using the anti-\(k_{t}\) algorithm [25] with a distance parameter \(R = 0.4\). Three-dimensional clusters of calorimeter cells called TopoClusters [26], calibrated using a local hadronic calibration [18], serve as inputs to the jet algorithm. The missing transverse momentum (with magnitude \({E_{\mathrm {T}}^{\mathrm {miss}}}\)) is computed from the combination of all reconstructed physics objects and the remaining calorimeter energy deposits not included in these objects  [27].

The ATLAS trigger system [19] consists of three levels; the first level (L1) is hardware-based while the second (L2) and third (Event Filter, EF) levels are software-based. The combination of L2 and the EF are referred to as the high-level trigger (HLT). The L1 trigger identifies regions-of-interest (RoI) using information from the calorimeters and the muon spectrometer. The delay between a beam crossing and the trigger decision (latency) is approximately 2 \(\upmu \)s at L1. The L2 system typically takes the RoIs produced by L1 as input and refines the quantities used for selection after taking into account the information from all subsystems. The latency at L2 is on average 40 ms, but can be as large as 100 ms at the highest instantaneous luminosities. At the EF level, algorithms similar to those run in the offline reconstruction are used to select interesting events with an average latency of about 1 s.

During 2012, the ATLAS detector was operated with a data-taking efficiency greater than \(95\,\%\). The highest peak luminosity obtained was \(8 \cdot 10^{33}~ \mathrm {cm} ^{-2} \mathrm {s}^{-1}\) at the end of 2012. The observed average number of pile-up interactions (meaning generally soft proton–proton interactions, superimposed on one hard proton–proton interaction) per bunch crossing in 2012 was 20.7. At the end of the data-taking period, the trigger system was routinely working with an average (peak) output rate of 700 Hz (1000 Hz).

2.2 Tau trigger operation

In 2012, a diverse set of tau triggers was implemented, using requirements on different final state configurations to maximize the sensitivity to a large range of physics processes. These triggers are listed in Table 1, along with the targeted physics processes and the associated kinematic requirements on the triggered objects. For the double hadronic triggers, in the lowest threshold version (29 and 20 GeV requirement on transverse momentum for the two \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)) two main criteria are applied: isolation at L1Footnote 2, and full tau identification at the HLT. The isolation requirement is dropped for the intermediate threshold version, and both criteria are dropped in favour of a looser (more than 95 % efficient), non-isolated trigger for the version with the highest thresholds.

As the typical rejection rates of \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)identification algorithms against the dominant multi-jet backgrounds are considerably smaller than those of electron or muon identification algorithms, \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)triggers must have considerably higher \(p_{\mathrm {T}}\) requirements in order to maintain manageable trigger rates. Therefore, most analyses using low-\(p_{\mathrm {T}}\) \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)in 2012 depend on the use of triggers which identify other objects. However, by combining tau trigger requirements with requirements on other objects, lower thresholds can be accommodated for the tau trigger objects as well as the other objects.

Figure 1 shows the tau trigger rates at L1 and the EF as a function of the instantaneous luminosity during the 8 \(\,\hbox {TeV}\) LHC operation. The trigger rates do not increase more than linearly with the luminosity, due the robust performance of the trigger algorithms under different pile-up conditions. The only exception is the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\) + \({E_{\mathrm {T}}^{\mathrm {miss}}}\) trigger, where the extra pile-up associated with the higher luminosity leads to a degradation of the resolution of the reconstructed event \({E_{\mathrm {T}}^{\mathrm {miss}}}\). At the highest instantaneous luminosities, the rates are affected by deadtime in the readout systems, leading to a general drop in the rates.

Fig. 1
figure 1

Tau trigger rates at a Level 1 and b Event Filter as a function of the instantaneous luminosity for \(\sqrt{s}=8\) \(\,\hbox {TeV}\). The triggers shown are described in Table 1, with the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)+\({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)being the rate for the lowest threshold trigger reported in the table. The rates for the higher threshold triggers are approximately three and five times lower at L1 and HLT, respectively, and are partially included in the rate of the lowest threshold item

2.3 Simulation and event samples

The optimization and measurement of tau performance requires simulated events. Events with \(Z/\gamma ^*\) and W boson production were generated using alpgen [28] interfaced to herwig [29] or Pythia6 [30] for fragmentation, hadronization and underlying-event (UE) modelling. In addition, \(Z \rightarrow \tau \tau \) and \(W \rightarrow \tau \nu \) events were generated using Pythia8 [31], and provide a larger statistical sample for the studies. For optimization at high \(p_{\mathrm {T}}\), \(Z' \rightarrow \tau \tau \) with \(Z'\) masses between 250 and 1250 \(\,\hbox {GeV}\) were generated with Pythia8. Top-quark-pair as well as single-top-quark events were generated with mc@nlo+herwig [32], with the exception of t-channel single-top production, where AcerMC+Pythia6 [33] was used. WZ and ZZ diboson events were generated with herwig, and WW events with alpgen+herwig. In all samples with \(\tau \) leptons, except for those simulated with Pythia8, Tauola [34] was used to model the \(\tau \) decays, and Photos [35] was used for soft QED radiative corrections to particle decays.

All events were produced using CTEQ6L1 [36] parton distribution functions (PDFs) except for the mc@nlo events, which used CT10 PDFs [37]. The UE simulation was tuned using collision data. Pythia8 events employed the AU2 tune [38], herwig events the AUET2 tune [39], while alpgen+Pythia6 used the Perugia2011C tune [40] and AcerMC+Pythia6 the AUET2B tune [41].

The response of the ATLAS detector was simulated using GEANT4 [42, 43] with the hadronic-shower model QGSP_BERT [44, 45] as baseline. Alternative models (FTFP_BERT [46] and QGSP) were used to estimate systematic uncertainties. Simulated events were overlaid with additional minimum-bias events generated with Pythia8 to account for the effect of multiple interactions occurring in the same and neighbouring bunch crossings (called pile-up). Prior to any analysis, the simulated events were reweighted such that the distribution of the number of pile-up interactions matched that in data. The simulated events were reconstructed with the same algorithm chain as used for collision data.

3 Reconstruction and identification of hadronic tau lepton decays

In the following, the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)reconstruction and identification at online and offline level are described. The trigger algorithms were optimized with respect to hadronic tau decays identified by the offline algorithms. This typically leads to online algorithms resembling their offline counterparts as closely as possible with the information available at a given trigger level. To reflect this, the details of the offline reconstruction and identification are described first, and then a discussion of the trigger algorithms follows, highlighting the differences between the two implementations.

3.1 Reconstruction

The \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)reconstruction algorithm is seeded by calorimeter energy deposits which have been reconstructed as individual jets. Such jets are formed using the anti-\(k_t\) algorithm with a distance parameter of \(R=0.4\), using calorimeter TopoClusters as inputs. To seed a \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate, a jet must fulfil the requirements of \(p_{\mathrm {T}}>10\) \(\,\hbox {GeV}\) and \({|\eta |}<2.5\). Events must have a reconstructed primary vertex with at least three associated tracks. In events with multiple primary vertex candidates, the primary vertex is chosen to be the one with the highest \(\Sigma p^2_\mathrm {T, tracks}\) value. In events with multiple simultaneous interactions, the chosen primary vertex does not always correspond to the vertex at which the tau lepton is produced. To reduce the effects of pile-up and increase reconstruction efficiency, the tau lepton production vertex is identified, amongst the previously reconstructed primary vertex candidates in the event.

The tau vertex (TV) association algorithm uses as input all tau candidate tracks which have \(p_{\mathrm {T}}>1\) \(\,\hbox {GeV}\), satisfy quality criteria based on the number of hits in the ID, and are in the region \(\Delta R<0.2\) around the jet seed direction; no impact parameter requirements are applied. The \(p_{T}\) of these tracks is summed and the primary vertex candidate to which the largest fraction of the \(p_{T}\) sum is matched to is chosen as the TV [47].

This vertex is used in the following to determine the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)direction, to associate tracks and to build the coordinate system in which identification variables are calculated. In \(Z \rightarrow \tau \tau \) events, the TV coincides with the highest \(\Sigma p^2_\mathrm {T,tracks}\) vertex (for the pile-up profile observed during 2012) roughly 90 % of the time. For physics analyses which require higher-\(p_{\mathrm {T}}\) objects, the two coincide in more than 99 % of all cases.

The \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)three-momentum is calculated by first computing \(\eta \) and \(\phi \) of the barycentre of the TopoClusters of the jet seed, calibrated at the LC scale, assuming a mass of zero for each constituent. The four-momenta of all clusters in the region \(\Delta R<0.2\) around the barycentre are recalculated using the TV coordinate system and summed, resulting in the momentum magnitude \(p^\mathrm {LC}\) and a \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)direction. The \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)mass is defined to be zero.

Tracks are associated with the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)if they are in the core region \(\Delta R<0.2\) around the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)direction and satisfy the following criteria: \(p_{\mathrm {T}}>1\) \(\,\hbox {GeV}\), at least two associated hits in the pixel layers of the inner detector, and at least seven hits in total in the pixel and the SCT layers. Furthermore, requirements are imposed on the distance of closest approach of the track to the TV in the transverse plane, \(|d_0|<1.0\) mm, and longitudinally, \(|z_0 \sin \theta |<1.5\) mm. When classifying a \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate as a function of its number of associated tracks, the selection listed above is used. Tracks in the isolation region \(0.2<\Delta R<0.4\) are used for the calculation of identification variables and are required to satisfy the same selection criteria.

A \(\pi ^{0}\) reconstruction algorithm was also developed. In a first step, the algorithm measures the number of reconstructed neutral pions (zero, one or two), \(N_{{\pi ^{0}}}\), in the core region, by looking at global tau features measured using strip layer and calorimeter quantities, and track momenta, combined in BDT algorithms. In a second step, the algorithm combines the kinematic information of tracks and of clusters likely stemming from \(\pi ^{0}\) decays. A candidate \(\pi ^{0}\) decay is composed of up to two clusters among those found in the core region of \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates. Cluster properties are used to assign a \(\pi ^{0}\) likeness score to each cluster found in the core region, after subtraction of the contributions from pile-up, the underlying event and electronic noise (estimated in the isolation region). Only those clusters with the highest scores are used, together with the reconstructed tracks in the core region of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate, to define the input variables for tau identification described in the next section.

3.2 Discrimination against jets

The reconstruction of \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates provides very little rejection against the jet background. Jets in which the dominant particleFootnote 3 is a quark or a gluon are referred to as quark-like and gluon-like jets, respectively. Quark-like jets are on average more collimated and have fewer tracks and thus the discrimination from \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)is less effective than for gluon-like jets. Rejection against jets is provided in a separate identification step using discriminating variables based on the tracks and TopoClusters (and cells linked to them) found in the core or isolation region around the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate direction. The calorimeter measurements provide information about the longitudinal and lateral shower shape and the \(\pi ^{0}\) content of tau hadronic decays.

The full list of discriminating variables used for tau identification is given below and is summarized in Table 2.

  • Central energy fraction (\(f_{\mathrm {cent}}\)): Fraction of transverse energy deposited in the region \(\Delta R < 0.1\) with respect to all energy deposited in the region \(\Delta R < 0.2\) around the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate calculated by summing the energy deposited in all cells belonging to TopoClusters with a barycentre in this region, calibrated at the EM energy scale. Biases due to pile-up contributions are removed using a correction based on the number of reconstructed primary vertices in the event.

  • Leading track momentum fraction (\(f_{\mathrm {track}}\)): The transverse momentum of the highest-\(p_{\mathrm {T}}\) charged particle in the core region of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate, divided by the transverse energy sum, calibrated at the EM energy scale, deposited in all cells belonging to TopoClusters in the core region. A correction depending on the number of reconstructed primary vertices in the event is applied to this fraction, making the resulting variable pile-up independent.

  • Track radius (\(R_{\mathrm {track}}\)): \(p_{\mathrm {T}}\)-weighted distance of the associated tracks to the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)direction, using all tracks in the core and isolation regions.

  • Leading track IP significance (\(S_{\mathrm {lead track}}\)): Transverse impact parameter of the highest-\(p_{\mathrm {T}}\) track in the core region, calculated with respect to the TV, divided by its estimated uncertainty.

  • Number of tracks in the isolation region (\(N^{\mathrm {iso}}_{\mathrm {track}}\)): Number of tracks associated with the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)in the region \(0.2<\Delta R<0.4\).

  • Maximum \(\Delta R\) (\(\Delta R_{\mathrm {Max}}\)): The maximum \(\Delta R\) between a track associated with the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate and the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)direction. Only tracks in the core region are considered.

  • Transverse flight path significance (\(S^{\mathrm {flight}}_{\mathrm {T}}\)): The decay length of the secondary vertex (vertex reconstructed from the tracks associated with the core region of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate) in the transverse plane, calculated with respect to the TV, divided by its estimated uncertainty. It is defined only for multi-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates.

  • Track mass (\(m_{\mathrm {track}}\)): Invariant mass calculated from the sum of the four-momentum of all tracks in the core and isolation regions, assuming a pion mass for each track.

  • Track-plus- \(\pi ^{0}\)-system mass (\(m_{\pi ^{0}+{\mathrm {track}}}\)): Invariant mass of the system composed of the tracks and \(\pi ^{0}\) mesons in the core region.

  • Number of \(\pi ^{0}\) mesons (\(N_{{\pi ^{0}} } \)): Number of \(\pi ^{0}\) mesons reconstructed in the core region.

  • Ratio of track-plus- \(\pi ^{0}\)-system \(p_{\mathrm {T}}\) (\( p_{\mathrm {T}}^{\pi ^{0} + {\mathrm {track}}} / p_{\mathrm {T}}\)): Ratio of the \(p_{\mathrm {T}}\) estimated using the track + \(\pi ^{0}\) information to the calorimeter-only measurement.

Table 2 Discriminating variables used as input to the tau identification algorithm at offline reconstruction and at trigger level, for 1-track and 3-track candidates. The bullets indicate whether a particular variable is used for a given selection. The \(\pi ^{0}\)-reconstruction-based variables, \(m_{\pi ^{0}+{\mathrm {track}}}\), \(N_{{\pi ^{0}}}\), \( p_{\mathrm {T}}^{\pi ^{0} + {\mathrm {track}}} / p_{\mathrm {T}}\) are not used in the trigger

The distributions of some of the important discriminating variables listed in Table 2 are shown in Figs. 2 and  3.

Fig. 2
figure 2

Signal and background distribution for the 1-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decay offline tau identification variables a \(f_{\mathrm {cent}}\) and b \(N^{\mathrm {iso}}_{\mathrm {track}}\). For signal distributions, 1-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays are matched to true generator-level \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)in simulated events, while the multi-jet events are obtained from the data

Fig. 3
figure 3

Signal and background distribution for the 3-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decay offline tau identification variables a \(R_{\mathrm {track}}\) and b \(m_{\pi ^{0}+{\mathrm {track}}}\). For signal distributions, 3-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays are matched to true generator-level \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)in simulated events, while the multi-jet events are obtained from data

Separate BDT algorithms are trained for 1-track and 3-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays using a combination of simulated tau leptons in Z, W and \(Z'\) decays. For the jet background, large collision data samples collected by jet triggers, referred from now on as the multi-jet data samples, are used. For the signal, only reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates matched to the true (i.e., generator-level) visible hadronic tau decay products in the region around \(\Delta R < 0.2\) with \(p^{\text {true}}_{\text {T,vis}} > 10\) \(\,\hbox {GeV}\) and \(|\eta ^{\text {true}}_{\text {vis}}| < 2.3\) are used. In the following, the signal efficiency is defined as the fraction of true visible hadronic tau decays with n charged decay products, which are reconstructed with n associated tracks and satisfy tau identification criteria. The background efficiency is the fraction of reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates with n associated tracks which satisfy tau identification criteria, measured in a background-dominated sample.

Fig. 4
figure 4

Offline tau identification efficiency dependence on the number of reconstructed interaction vertices, for a 1-track and b 3-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays matched to true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)(with corresponding number of charged decay products) from SM and exotic processes in simulated data. Three working points, corresponding to different tau identification efficiency values, are shown

Three working points, labelled tight, medium and loose, are provided, corresponding to different tau identification efficiency values. Their signal efficiency values (defined with respect to 1-track or 3-track reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates matched to true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)) can be seen in Fig. 4. The requirements on the BDT score are chosen such that the resulting efficiency is independent of the true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\) \(p_{\mathrm {T}}\). Due to the choice of input variables, the tau identification also shows stability with respect to the pile-up conditions as shown in Fig. 4. The performance of the tau identification algorithm in terms of the inverse background efficiency versus the signal efficiency is shown in Fig. 5. At low transverse momentum of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates, 40 % signal efficiency for an inverse background efficiency of 60 is achieved. The signal efficiency saturation point, visible in these curves, stems from the reconstruction efficiency for a true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with one or three charged decay products to be reconstructed as a 1-track or 3-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate. The main sources of inefficiency are track reconstruction efficiency due to hadronic interactions and migration of the number of reconstructed tracks due to conversions or underlying-event tracks being erroneously associated with the tau candidate.

Fig. 5
figure 5

Inverse background efficiency versus signal efficiency for the offline tau identification, for a a low-\(p_{\mathrm {T}}\) and b a high-\(p_{\mathrm {T}}\) \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)range. Simulation samples for signal include a mixture of Z, W and \(Z'\) production processes, while data from multi-jet events is used for background. The red markers correspond to the three working points mentioned in the text. The signal efficiency shown corresponds to the total efficiency of \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays to be reconstructed as 1-track or 3-track and pass tau identification selection

3.3 Tau trigger implementation

The tau reconstruction at the trigger level has differences with respect to its offline counterpart due to the technical limitations of the trigger system. At L1, no inner detector track reconstruction is available, and the full calorimeter granularity cannot be accessed. Latency limits at L2 prevent the use of the TopoCluster algorithm, and only allow the candidate reconstruction to be performed within the given RoI. At the EF, the same tau reconstruction and identification methods as offline are used, except for the \(\pi ^{0}\) reconstruction. In this section, the details of the tau trigger reconstruction algorithm are described.

Level 1 At L1, the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates are selected using calorimeter energy deposits. Two calorimeter regions are defined by the tau trigger for each candidate, using trigger towers in both the EM and HAD calorimeters: the core region, and an isolation region around this core. The trigger towers have a granularity of \(\Delta \eta \times \Delta \phi =0.1\times 0.1\) with a coverage of \({|\eta |}< 2.5\). The core region is defined as a square of \(2\times 2\) trigger towers, corresponding to \(0.2\times 0.2\) in \(\Delta \eta \times \Delta \phi \) space. The \(E_{\mathrm {T}}\) of a \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate at L1 is taken as the sum of the transverse energy in the two most energetic neighbouring central towers in the EM calorimeter core region, and in the \(2\times 2\) towers in the HAD calorimeter, all calibrated at the EM scale. For each \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate, the EM isolation is calculated as the transverse energy deposited in the annulus between \(0.2\times 0.2\) and \(0.4\times 0.4\) in the EM calorimeter.

To suppress background events and thus reduce trigger rates, an EM isolation energy of less than 4 \(\,\hbox {GeV}\) is required for the lowest \(E_{\mathrm {T}}\) threshold at L1. Hardware limitations prevent the use of an \(E_{\mathrm {T}}\)-dependent selection. This requirement reduces the efficiency of \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)events by less than 2 % over most of the kinematic range. Larger efficiency losses occur for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)events at high \(E_{\mathrm {T}}\) values; those are recovered through the use of triggers with higher \(E_{\mathrm {T}}\) thresholds but without any isolation requirements.

The energy resolution at L1 is significantly lower than at the offline level. This is due to the fact that all cells in a trigger tower are combined without the use of sophisticated clustering algorithms and without \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)-specific energy calibrations. Also, the coarse energy and geometrical position granularity limits the precision of the measurement. These effects lead to a significant signal efficiency loss for low-\(E_{\mathrm {T}}\) \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates.

Level 2 At L2, \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate RoIs from L1 are used as seeds to reconstruct both the calorimeter- and tracking-based observables associated with each \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate. The events are then selected based on an identification algorithm that uses these observables. The calorimeter observables associated with the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates are calculated using calorimeter cells, where the electronic and pile-up noise are subtracted in the energy calibration. The centre of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy deposit is taken as the energy-weighted sum of the cells collected in the region \(\Delta R<0.4\) around the L1 seed. The transverse energy of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)is calculated using only the cells in the region \(\Delta R<0.2\) around its centre.

To calculate the tracking-based observables, a fast tracking algorithm [48] is applied, using only hits from the pixel and SCT tracking layers. Only tracks satisfying \(p_{\mathrm {T}}> 1.5\) \(\,\hbox {GeV}\) and located in the region \(\Delta R<0.3\) around the L2 calorimeter \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)direction are used. The tracking efficiency with respect to offline reaches a plateau of 99 % at 2 \(\,\hbox {GeV}\) (with an efficiency of about 98 % at 1.5 \(\,\hbox {GeV}\)). The fast tracking algorithm required an average of 37 ms to run at the highest pile-up conditions at peak luminosity in 2012 (approximately forty pile-up interactions).

As there is no vertex information available at this stage, an alternative approach is used to reject tracks coming from pile-up interactions. A requirement is placed on the \(\Delta z_{0}\) between a candidate track and the highest-\(p_{\mathrm {T}}\) track inside the RoI. The distribution of \(\Delta z_{0}\) is shown in Fig. 6 for simulated \(Z \rightarrow \tau \tau \) events with an average of eight interactions per bunch crossing. High values of \(\Delta z_{0}\) typically correspond to pile-up tracks while the central peak corresponds to the main interaction tracks.

The \(\Delta z_{0}\) distribution is fit to the sum of a Breit–Wigner function to describe the central peak and a Gaussian function to describe the broad distribution from tracks in pile-up events. The half-width of the Breit–Wigner \(\sigma \)=0.32 mm is taken as the point where 68 % of the signal events are included in the central peak. A dependence of the trigger variables on pile-up conditions is minimized by considering only tracks within \(-2\) mm \(<\Delta z_{0}<\) 2 mm and \(\Delta R<0.1\) with respect to the highest-\(p_{\mathrm {T}}\) track.

Track isolation requirements are applied to \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates to increase background rejection. For multi-track candidates (candidates with two or three associated tracks, defined to be as inclusive as possible with respect to their offline counterpart), the ratio of the sum of the track \(p_{\mathrm {T}}\) in \(0.1<\Delta R<0.3\) to the sum of the track \(p_{\mathrm {T}}\) in \(\Delta R<0.1\) is required to be lower than 0.1. Any 1-track candidate with a reconstructed track in the isolation region is rejected.

In the last step, identification variables combining calorimeter and track information are built as described in Sect. 3.2. The calorimeter-based isolation variable \(f_{\mathrm {cent}}\) uses an expanded cone size of \(\Delta R < 0.4\) without the pile-up correction term to estimate the fraction of transverse energy deposited in the region \(\Delta R < 0.1\) around the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate. The variables \(f_{\mathrm {track}}\) and \(R_{\mathrm {track}}\), measuring respectively the ratio of the transverse momentum of the leading \(p_{\mathrm {T}}\) track to the total transverse energy (calibrated at the EM energy scale) and the \(p_{\mathrm {T}}\)-weighted distance of the associated tracks to the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)direction, are calculated using selected tracks in the region \(\Delta R < 0.3\) around the highest-\(p_{\mathrm {T}}\) track. Cuts on the chosen identification variables are optimized to provide an inverse background efficiency of roughly ten while keeping the signal efficiency as high as possible (approximately 90 % with respect to the offline medium tau identification).

Fig. 6
figure 6

Distribution of \(\Delta z_{0}\) for the tau trigger at L2 in simulated \(Z \rightarrow \tau \tau \) events with an average of eight interactions per bunch crossing. The wide Gaussian distribution corresponds to pile-up tracks while the central peak, displayed in the upper-right corner, corresponds to the main interaction tracks. A Breit–Wigner function is fitted to the central peak and 68 % of the signal events are found within a distance \(\sigma \) = 0.32 mm from the peak

Event Filter At the EF level, the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)reconstruction is very similar to the offline version. First, the TopoCluster reconstruction and calibration algorithms are run within the RoI. Then, track reconstruction inside the RoI is performed using the EF tracking algorithm. In the last step, the full offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)reconstruction algorithm is used. The EF tracking is almost 100 % efficient over the entire \(p_{\mathrm {T}}\) range with respect to the offline reconstructed tracks. It is, however, considerably slower than the L2 fast tracking algorithm, requiring about 200 ms per RoI under severe pile-up conditions (forty pile-up interactions). The TopoClustering algorithms need only about 15 ms.

The \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\) candidate four-momentum and input variables to the EF tau identification are then calculated. The main difference with respect to the offline tau reconstruction is that \(\pi ^{0}\)-reconstruction-based input variables (\(m_{\pi ^{0}+{\mathrm {track}}}\), \(N_{{\pi ^{0}}}\) and \( p_{\mathrm {T}}^{\pi ^{0} + {\mathrm {track}}} / p_{\mathrm {T}}\)) are not used; the methodology to compute these variables had not yet been developed when the trigger was implemented. Furthermore, no pile-up correction is applied to the input variables at trigger level.

Since full-event vertex reconstruction is not available at trigger level (vertices are only formed using the tracks in a given RoI), the selection requirements applied to the input tracks are also different with respect to the offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)reconstruction. Similarly to L2, the \(\Delta z_{0}\) requirement for tracks is computed with respect to the leading track, and loosened to 1.5 mm with respect to the offline requirement. The \(\Delta d_{0}\) requirement is calculated with respect to the vertex found inside of the RoI, and is loosened to 2 mm.

A BDT with the input variables listed in Table 2 is used to suppress the backgrounds from jets misidentified as \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\). The BDT was trained on 1- and 3-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates with simulated Z, W and \(Z'\) events for the signal and data multi-jet samples for the background, respectively. Only events passing an L1 tau trigger matched with an offline reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with \(p_{\mathrm {T}}>15\) \(\,\hbox {GeV}\) and \({|\eta |}<2.2\) are used, where the medium identification is required for the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates. For the signal, in addition, a geometrical matching to a true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)is required. The performance of the EF tau trigger is presented in Fig. 7. The signal efficiency is defined with respect to offline reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates matched at generator level, and the inverse background efficiency is calculated in a multi-jet sample. The working points are chosen to obtain a signal efficiency of 85 and 80 % with respect to the offline medium candidates for 1-track and multi-track candidates respectively, where the inverse background efficiency is of the order of 200 for the multi-jet sample.

Fig. 7
figure 7

Inverse background efficiency versus signal efficiency for the tau trigger at the EF level, for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates which have satisfied the L1 requirements. The signal efficiency is defined with respect to offline medium tau identification \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates matched at generator level, and the inverse background efficiency is calculated in a multi-jet sample

3.4 Discrimination against electrons and muons

Additional dedicated algorithms are used to discriminate \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)from electrons and muons. These algorithms are only used offline.

Electron veto The characteristic signature of 1-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)can be mimicked by electrons. This creates a significant background contribution after all the jet-related backgrounds are suppressed via kinematic, topological and \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)identification criteria. Despite the similarities of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)and electron signatures, there are several properties that can be used to discriminate between them: transition radiation, which is more likely to be emitted by an electron and causes a higher ratio \(f_{\mathrm {HT}}\) of high-threshold to low-threshold track hits in the TRT for an electron than for a pion; the angular distance of the track from the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)calorimeter-based direction; the ratio \(f_{\mathrm {EM}}\) of energy deposited in the EM calorimeter to energy deposited in the EM and HAD calorimeters; the amount of energy leaking into the hadronic calorimeter (longitudinal shower information) and the ratio of energy deposited in the region \(0.1<\Delta R <0.2\) to the total core region \(\Delta R<0.2\) (transverse shower information). The distributions for two of the most powerful discriminating variables are shown in Fig. 8. These properties are used to define a \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)identification algorithm specialized in the rejection of electrons misidentified as hadronically decaying tau leptons, using a BDT. The performance of this electron veto algorithm is shown in Fig. 9. Slightly different sets of variables are used in different \(\eta \) regions. One of the reasons for this is that the variable associated with transition radiation (the leading track’s ratio of high-threshold TRT hits to low-threshold TRT hits) is not available for \(|\eta |>\) 2.0. Three working points, labelled tight, medium and loose are chosen to yield signal efficiencies of 75, 85, and 95 %, respectively.

Fig. 8
figure 8

Signal and background distribution for two of the electron veto variables, a \(f_{\mathrm {HT}}\) and b \(f_{\mathrm {EM}}\). Candidate 1-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays are required to not overlap with a reconstructed electron candidate which passes tight electron identification [23]. For signal distributions, 1-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays are matched to true generator-level \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)in simulated \(Z\rightarrow \tau \tau \) events, while the electron contribution is obtained from simulated \(Z\rightarrow ee\) events where 1-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays are matched to true generator-level electrons

Fig. 9
figure 9

Electron veto inverse background efficiency versus signal efficiency in simulated samples, for 1-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates. The background efficiency is determined using simulated \(Z \rightarrow ee\) events

Muon veto Tau candidates corresponding to muons can in general be discarded based on the standard muon identification algorithms [24]. The remaining contamination level can typically be reduced to a negligible level by a cut-based selection using the following characteristics. Muons are unlikely to deposit enough energy in the calorimeters to be reconstructed as \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates. However, when a sufficiently energetic cluster in the calorimeter is associated with a muon, the muon track and the calorimeter cluster together may be misidentified as a \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\). Muons which deposit a large amount of energy in the calorimeter and therefore fail muon spectrometer reconstruction are characterized by a low electromagnetic energy fraction and a large ratio of track-\(p_{\mathrm {T}}\) to \(E_{\mathrm {T}}\) deposited in the calorimeter. Low-momentum muons which stop in the calorimeter and overlap with calorimeter deposits of different origin are characterized by a large electromagnetic energy fraction and a low \(p_{\mathrm {T}}\)-to-\(E_{\mathrm {T}}\) ratio. A simple cut-based selection based on these two variables reduces the muon contamination to a negligible level. The resulting efficiency is better than 96 % for true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\), with a reduction of muons misidentified as \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)of about 40 %. However, the performance can vary depending on the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)and muon identification levels.

4 Efficiency measurements using Z tag-and-probe data

To perform physics analyses it is important to measure the efficiency of the reconstruction and identification algorithms used online and offline with collision data. For the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)signal, this is done on a sample enriched in \(Z \rightarrow \tau \tau \) events. For electrons misidentified as a tau signal (after applying the electron veto) this is done on a sample enriched in \(Z \rightarrow ee\) events.

The chosen tag-and-probe approach consists of selecting events triggered by the presence of a lepton (tag) and containing a hadronically decaying tau lepton candidate (probe) in the final state and extracting the efficiencies directly from the number of reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)before and after tau identification algorithms are applied. In practice, it is impossible to obtain a pure sample of hadronically decaying tau leptons, or electrons misidentified as a tau signal, and therefore backgrounds have to be taken into account. This is described in the following sections.

4.1 Offline tau identification efficiency measurement

To estimate the number of background events for the purpose of tau identification efficiency measurements, a variable with high separation power, which is modelled well for simulated \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)decays is chosen: the sum of the number of core and outer tracks associated to the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate. Outer tracks in \(0.2 < \Delta R < 0.6\) are only considered if they fulfill the requirement \(D^\mathrm {outer} = min([~p_{\mathrm {T}}^\mathrm {core} / p_{\mathrm {T}}^\mathrm {outer}~] \cdot \Delta R(\mathrm {core,outer})) <\) 4, where \(p_{\mathrm {T}}^\mathrm {core}\) refers to any track in the core region, and \(\Delta R(\mathrm {core,outer})\) refers to the distance between the candidate outer track and any track in the core region. This requirement suppresses the contribution of outer tracks from underlying and pile-up events, due to requirements on the relative momentum and separation of the tracks. This allows the signal track multiplicity to retain the same structure as the core track multiplicity distribution. For backgrounds from multi-jet events, the track multiplicity is increased by the addition of tracks with significant momentum in the outer cone. The requirement on \(D^\mathrm {outer}\) was chosen to offer optimal signal to background separation. A fit is then performed using the expected distributions of this variable for both signal and background to extract the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)signal. This fit is performed for each exclusive tau identification working point, corresponding to: candidates failing the loose requirement, candidates satisfying the loose requirement but failing the medium requirement, candidates satisfying the medium requirement but failing the tight requirement and candidates satisfying the tight requirement.

4.1.1 Event selection

\(Z\rightarrow {\tau _{\mathrm {lep}}}\tau _{\mathrm {had}}\) events are selected by a triggered muon or electron coming from the leptonic decay of a tau lepton, and the hadronically decaying tau lepton is then searched for in the rest of the event, considered as the probe for the tau identification performance measurement. These events are triggered by a single-muon or a single-electron trigger requiring one isolated trigger muon or electron with a \(p_\mathrm{T}\) of at least 24 \(\,\hbox {GeV}\).

Offline, muons and electrons with \(p_\mathrm{T}>26\) \(\,\hbox {GeV}\) are thereafter selected, representing the tag objects. Additional track and calorimeter isolation requirements are applied to the muon and electron. Identified muons are required to have \(|\eta |<2.4\). Identified electrons are required to have \(|\eta |<1.37\) or \(1.52<|\eta |<2.47\), therefore excluding the poorly instrumented region at the interface between the barrel and endcap calorimeters. In addition to the requirement of exactly one isolated muon or electron (\(\ell \)), a \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate is selected in the kinematic range \(p_\mathrm{T} > 15\) \(\,\hbox {GeV}\) and \(|\eta |<2.5\), requiring one or three associated tracks in the core region and an absolute electric charge of one and no geometrical overlap with muons with \(p_{\mathrm {T}}{} > 4\) \(\,\hbox {GeV}\) or with electrons with \(p_{\mathrm {T}}{} > 15\) \(\,\hbox {GeV}\) of loose or medium quality (depending on \(\eta \)). For \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with one associated track, a muon veto and a medium electron veto is applied. In addition to this, a very loose requirement on the tau identification BDT score is made which strongly suppresses jets while being more than 99 % efficient for \(Z \rightarrow \tau \tau \) signal. The tag and the probe objects are required to have opposite-sign electric charges (OS).

Additional requirements are made in order to suppress \((Z \rightarrow \ell \ell )\) + jets and \((W\rightarrow \ell \nu _{\ell })\) + jets events:

  • On the invariant mass calculated from the lepton and the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)four-momenta (\(m_\mathrm{vis}(\ell ,{{\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}})\)): for \(p_{\mathrm {T}}^{{{\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}}}<20 \,\hbox {GeV}\), \(45 \,\hbox {GeV}<m_\mathrm{vis}( \ell ,{{\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}})< 80 \,\hbox {GeV}\). Otherwise, for the \(\mu \) channel, \(50 \,\hbox {GeV}<m_\mathrm{vis}(\mu ,{{\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}})< 85 \,\hbox {GeV}\), and for the e channel: \(50 \,\hbox {GeV}<m_\mathrm{vis}( e ,{{\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}})< 80 \,\hbox {GeV}\). For the signal, this variable peaks in these regions.

  • On the transverse mass of the lepton and \({E_{\mathrm {T}}^{\mathrm {miss}}}\) system (\(m_\mathrm{T}=\sqrt{2p^{\ell }_\mathrm{T} \cdot E^\mathrm{miss}_\mathrm{T}(1-\cos \Delta \phi (\ell ,E^\mathrm{miss}_\mathrm{T}))}\)): \(m_\mathrm{T} <\) 50 GeV. For most backgrounds (e.g. \((W\rightarrow ~\ell \nu _{\ell })\) + jets), this variable peaks at larger values.

  • On the distance in the azimuthal plane between the lepton and \({E_{\mathrm {T}}^{\mathrm {miss}}}\) (neutrinos) and between the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)and \({E_{\mathrm {T}}^{\mathrm {miss}}}\) (\( \Sigma \cos \Delta \phi =\cos \Delta \phi (\ell , E^\mathrm{miss}_\mathrm{T})+\cos \Delta \phi ({{\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}}, E^\mathrm{miss}_\mathrm{T})\)): \( \Sigma \cos \Delta \phi > -0.15\). For the signal, this variable tends to peak at zero, indicating that the neutrinos point mainly in the direction of one of the two leptons from Z decay products. For W + jets background events, the value is typically negative, indicating that the neutrino points away from the two lepton candidates.

4.1.2 Background estimates and templates

The signal track multiplicity distribution is modelled using simulated \(Z\rightarrow \tau _{\mathrm {lep}}\tau _{\mathrm {had}}\) events. Only reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)matched to a true hadronic tau decay are considered.

A single template is used to model the background from quark- and gluon-initiated jets that are misidentified as hadronic tau decays. The background is mainly composed of multi-jet and W+jets events with a minor contribution from Z+jets events. The template is constructed starting from a enriched multi-jet control region in data that uses the full signal region selection but requires that the tag and probe objects have same-sign charges (SS). The contributions from W+jets and Z+jets in the SS control region are subtracted. The template is then scaled by the ratio of OS / SS multi-jet events, measured in a control region which inverts the very loose identification requirement of the signal region. Finally, the OS contributions from W+jets and Z+jets are added to complete the template. The Z+jets contribution is estimated using simulated samples. The shape of the W+jets contribution is estimated from a high-purity W+jets control region, defined by removing the \(m_{\mathrm {T}}\) requirement and inverting the requirement on \( \Sigma \cos \Delta \phi \). The normalization of the W+jets contribution is estimated using simulation.

An additional background shape is used to take into account the contamination due to misidentified electrons or muons. This small background contribution (stemming mainly from \(Z\rightarrow \ell \ell \) events) is modelled by taking the shape predicted by simulation using candidates which are not matched to true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\) in events of type \(Z\rightarrow \tau _{\mathrm {lep}}\tau _{\mathrm {had}}\), \(t\bar{t}\), diboson, \(Z\rightarrow ee,\mu \mu \) where the reconstructed tau candidate probe is matched to a electron or muon. For the fit, the contribution of these backgrounds is fixed to the value predicted by the simulation, which is typically less than 5 % of the total signal yield.

To measure both the 1-track and 3-tracks \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\) efficiencies, a fit of the data to the model (signal plus background) is performed, using two separate signal templates. The signal templates are obtained by requiring exactly one or three tracks reconstructed in the core region of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate. To improve the fit stability in the background-dominated region where the tau candidates fail the loose requirements, the ratio of the 1-track to 3-track normalization is fixed to the value predicted by the simulation. For other exclusive regions, the ratio is allowed to vary during the fit.

In the fit to extract the efficiencies for real tau leptons passing different levels of identification, the ratio of jet to other \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates is determined in a preselection step (where no identification is required) and then extrapolated to regions where identification is required by using jet misidentification rates determined in an independent data sample.

4.1.3 Results

Figure 10 shows an example of the track multiplicity distribution after the tag-and-probe selection, before and after applying the tau identification requirements, with the results of the fit performed. The peaks in the one- and three-track bins are due to the signal contribution. These are visible before any identification requirements are applied, and become considerably more prominent after identification requirements are applied, due to the large amount of background rejection provided by the identification algorithm. To account for the small differences between data and simulations, correction factors, defined as the ratio of the efficiency in data to the efficiency in simulation for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\) signal to pass a certain level of identification, are derived. Their values are compatible with one, except for the tight 1-track working point, where the correction factor is about 0.9.

Fig. 10
figure 10

Template fit result in the muon channel, inclusive in \(\eta \) and \(p_{\mathrm {T}}\) for \(p_{\mathrm {T}}>20\)  \(\,\hbox {GeV}\) for the offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates a before the requirement of tau identification, and b fulfilling the medium tau identification requirement

Fig. 11
figure 11

Correction factors needed to bring the offline tau identification efficiency in simulation to the level observed in data, for all tau identification working points as a function of \(\eta \). The combinations of the muon and electron channels are also shown, and the results are displayed separately for a 1-track and b 3-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates with \(p_{\mathrm {T}}{} > 20\) \(\,\hbox {GeV}\). The combined systematic and statistical uncertainties are shown

Results from the electron- and muon-tag analysis are combined to improve the precision of the correction factors, shown in Fig. 11. No significant dependency on the \(p_{\mathrm {T}}\) of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)is observed and hence the results are provided separately only for the barrel (\(|\eta | < 1.5\)) and the endcap (\(1.5 < |\eta | < 2.5\)) region, and for one and three associated tracks. Uncertainties depend slightly on the tau identification level and kinematic quantities. In Table 3, the most important systematic uncertainties for the working point used by most analyses, medium tau identification, are shown, together with the total statistical and systematic uncertainty. Uncertainties due to the underlying event (UE) are the dominant ones for the signal template, and are estimated by comparing alpgen-Herwig and Pythia simulations. The shower model and the amount of detector material are also varied and included in the number reported in Table 3. The W+jets shape uncertainty accounts for differences between the W+jets shape in the signal and control regions and is derived from comparisons to simulated W+jets events. The jet background fraction uncertainty accounts for the effect of propagating the statistical uncertainty on the jet misidentification rates.

The results apply to \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates with \(p_{\mathrm {T}}{} > 20\) \(\,\hbox {GeV}\). For \(p_{\mathrm {T}}{} < \) 20 \(\,\hbox {GeV}\), uncertainties increase to a maximum of 15 % for inclusive \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates. For \(p_{\mathrm {T}}{} >\) 100 GeV, there are no abundant sources of hadronic tau decays to allow for an efficiency measurement. Previous studies using high-\(p_{\mathrm {T}}{}\) dijet events indicate that there is no degradation in the modelling of tau identification in this \(p_{\mathrm {T}}{}\) range, within the statistical uncertainty of the measurement [14].

Table 3 Dominant uncertainties on the medium tau identification efficiency correction factors estimated with the Z tag-and-probe method, and the total uncertainty, which combines systematic and statistical uncertainties. These uncertainties apply to \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates with \(p_{\mathrm {T}}\) \(>\) 20 GeV

4.2 Trigger efficiency measurement

The tau trigger efficiency is measured with \(Z \rightarrow \tau \tau \) events using tag-and-probe selection similar to the one described in Sect. 4.1. The only difference is that the efficiency is measured with respect to identified offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates and thus, offline tau identification selection criteria are applied during the event selection. Only the muon channel is considered, as the background contamination is smaller than in the electron channel. The statistical uncertainty improvements that could be obtained by the addition of the electron channel are offset by the larger systematic uncertainties associated with this channel. The systematic uncertainties are also different from those in the offline identification measurement, since the purity after identification is already high. The systematics are dominated by the uncertainties on the modelling of the kinematics of the background events, rather than the total normalization, as is the case for the offline identification measurement.

The dominant background contribution is due to W+jets and multi-jet events, where a jet is misidentified as a \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\). These backgrounds are estimated using a method similar to the one described in Sect. 4.1.2. The same multi-jet and W+jets control regions are used. The shape of other backgrounds is taken from simulation but the normalizations of the dominant backgrounds are estimated from data control regions. The contribution of top quark events is normalized in a control region requiring one jet originating from a b-quark. Z+jets events with leptonic Z decays and one of the additional jets being misidentified as \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)are normalized by measuring this misidentification rate in a control region with two identified oppositely charged same-flavour leptons.

In total, more than 60,000 events are collected, with a purity of about 80 % when the offline medium tau identification requirement is applied. With the addition of the tau trigger requirement, the purity increases to about 88 %. Most of the backgrounds accumulate in the region \(p_{\mathrm {T}}<30\) \(\,\hbox {GeV}\).

Figure 12 shows the measured tau trigger efficiency for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates identified by the offline medium tau identification as functions of the offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)transverse energy and the number of primary vertices in the event, for each level of the trigger. The tau trigger considered has calorimetric isolation and a \(p_{\mathrm {T}}\) threshold of 11 \(\,\hbox {GeV}\) at L1, a 20 \(\,\hbox {GeV}\) requirement on \(p_{\mathrm {T}}\), the number of tracks restricted to three or less, and medium selection on the BDT score at EF. The efficiency depends minimally on \(p_{\mathrm {T}}\) for \(p_{\mathrm {T}}>35\) \(\,\hbox {GeV}\) or on the pile-up conditions. The measured tau trigger efficiency is compared to simulation in Fig. 13; the efficiency is shown to be modelled well in simulation. Correction factors, as defined in Sect. 4.1, are derived from this measurement. The correction factors are in general compatible with unity, except for the region \(p_{\mathrm {T}}<40\) \(\,\hbox {GeV}\) where a difference of a few percent is observed.

Fig. 12
figure 12

The tau trigger efficiency for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates identified by the offline medium tau identification, as a function of a the offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)transverse energy and b the number of primary vertices. The error bars correspond to the statistical uncertainty in the efficiency

In the \(p_{\mathrm {T}}\) range from 30 to 50 \(\,\hbox {GeV}\), the uncertainty on the correction factors is about 2 % but increases to about 8 % for \(p_{\mathrm {T}}=100\) \(\,\hbox {GeV}\). The uncertainty is also sizeable in the region \(p_{\mathrm {T}}<30\) \(\,\hbox {GeV}\), where the background contamination is the largest.

Fig. 13
figure 13

The measured tau trigger efficiency in data and simulation, for the offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates passing the medium tau identification, as a function of offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)transverse energy. The expected background contribution has been subtracted from the data. The uncertainty band on the ratio reflects the statistical uncertainties associated with data and simulation and the systematic uncertainty associated with the background subtraction in data

4.3 Electron veto efficiency measurement

To measure the efficiency for electrons reconstructed as \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)to pass the electron veto in data, a tag-and-probe analysis singles out a pure sample of \(Z \rightarrow ee\) events, as illustrated in Fig. 14a. The measurement uses probe 1-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates in the opposite hemisphere to the identified tag electron. The tag electron is required to fulfil \(p_{\mathrm {T}}^{\mathrm {tag}} > 35\) GeV in order to suppress backgrounds from \(Z \rightarrow \tau \tau \) events. The probe is required not to overlap geometrically with an identified electron, e.g. in the case of Fig. 14 a loose electron identification is used. Different veto algorithms are tested in combination with different levels of jet discrimination, and the effects estimated. Efficiencies are extracted directly from the number of reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)before and after identification, in bins of \(\eta \) of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate, after subtracting the background modelled by simulation (normalized to data in dedicated control regions). The shape and normalization of the multi-jet background distribution for the \(\eta \) of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)are estimated using events with SS tag electron and probe \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)in data after subtracting backgrounds in the SS region using simulation. To estimate the \(W\rightarrow e\nu \), \(Z\rightarrow \tau \tau \), and \(t\bar{t}\) backgrounds, the shape of this distribution is obtained from simulation but normalized to dedicated data control regions for each background.

Differences in the modelling of the electron veto algorithm’s performance in simulation compared to data are parameterized as correction factors in bins of \(\eta \) of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidate, by comparing distributions similar to the one shown in Fig. 14.

Fig. 14
figure 14

a Visible mass of electron–positron pairs for the offline electron veto efficiency measurement, after tag-and-probe selection, where the probe lepton passes medium tau identification and does not overlap with loose electrons, before the electron veto is applied. b \(\eta \) distribution for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates (electrons misidentified as hadronic tau decays) after applying a loose electron veto. Uncertainties shown are only statistical

Uncertainties on the correction factors (which are typically close to unity) are \(\eta \)-dependent and amount to about 10 % for the loose electron veto and get larger for the medium and tight electron veto working points, mainly driven by statistical uncertainties. A summary of the main uncertainties for the working point shown in Fig. 14 is provided in Table 4.

Table 4 Dominant uncertainties on the loose electron veto efficiency correction factors estimated with the Z tag-and-probe method. The range of the uncertainties reflects their variation with \(\eta \)

5 Calibration of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy

The \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy calibration is done in several steps. First, a calibration described in Sect. 5.1 and derived from simulation brings the tau energy scale (TES) into agreement with the true energy scale at the level of a few percent and removes any significant dependencies of the energy scale on the pseudorapidity, energy, pile-up conditions and track multiplicity. Then, additional small corrections to the TES are derived using one of two independent data-driven methods described in Sect. 5.2. Which of the two methods is used depends on whether for a given study the agreement between reconstructed and true TES or the modelling of the TES in simulation is more important.

5.1 Offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy calibration

The clusters associated with the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)reconstruction are calibrated at the LC scale. For anti-\(k_t\) jets with a distance parameter \(R=0.4\), this calibration accounts for the non-compensating nature of the ATLAS calorimeters and for energy deposited outside the reconstructed clusters and in non-sensitive regions of the calorimeters. However, it is neither optimized for the cone size used to measure the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)momentum (\(\Delta R=0.2\)) nor for the specific mix of hadrons observed in tau decays; and it does not correct for the underlying event or for pile-up contributions. Thus an additional correction is needed to obtain an energy scale which is in agreement with the true visible energy scale, thereby also improving the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy resolution.

This correction (also referred to as a response curve) is computed as a function of \(E^\tau _\mathrm {LC}\) using \(Z\rightarrow \tau \tau \), \(W \rightarrow \tau \nu \) and \(Z' \rightarrow \tau \tau \) events simulated with Pythia8. Only \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates with reconstructed \(E_{\mathrm {T}}> 15\) \(\,\hbox {GeV}\) and \({|\eta |}<2.4\) matched to a true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with \(E^\mathrm {true}_{\mathrm {T,vis}} >10\) \(\,\hbox {GeV}\) are considered. Additionally, they are required to satisfy medium tau identification criteria and to have a distance \(\Delta R>0.5\) to other reconstructed jets. The response is defined as the ratio of the reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy at the LC scale \(E^\tau _\mathrm {LC}\) to the true visible energy \(E^\mathrm {true}_\mathrm {vis}\).

The calibration is performed in two steps: first, the response curve is computed; then, additional small corrections for the pseudorapidity bias and for pile-up effects are derived.

The response curve is evaluated in intervals of \(E^\mathrm {true}_{\mathrm {vis}}\) and of the absolute value of the reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)pseudorapidity for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates with one or more tracks. In each interval, the distribution of this ratio is fitted with a Gaussian function to determine the mean value. This mean value as a function of the average \(E^\tau _\mathrm {LC}\) in a given interval is then fitted with an empirically derived functional form. The resulting functions are shown in Fig. 15.

Fig. 15
figure 15

Offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy response curves as a function of the reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy \(E^\tau _\mathrm {LC}\) for hadronic tau decays with a one and b more than one associated tracks. One curve per pseudorapidity region \(|\eta ^\mathrm {LC}|\) is shown. The region where markers are shown corresponds approximately to a transverse energy \(E^\tau _\mathrm {T,LC}>15\) \(\,\hbox {GeV}\). For very low and very high energies, the response curves are assumed to be constant. Uncertainties are statistical only

After using this response curve to calibrate hadronically decaying tau leptons their reconstructed mean energy is within 2 % of the final scale, which is set using two additional small corrections. First, a pseudorapidity correction is applied, which is necessary to counter a bias due to underestimated reconstructed cluster energies in poorly instrumented regions. The correction depends only on \(|\eta ^\mathrm {LC}|\) and is smaller than 0.01 units in the transition region between the barrel and endcap electromagnetic calorimeters and negligible elsewhere, leading to the final reconstructed pseudorapidity \(\eta ^\mathrm {rec}=\eta ^\mathrm {LC} - \eta ^\mathrm {bias}\).

Pile-up causes response variations of typically a few percent. This is corrected by subtracting an amount of energy which is proportional to the number of reconstructed proton–proton interaction vertices \(n_\mathrm {vtx}\) in a given event. The parameter describing the proportionality is derived for different regions of \(|\eta ^\mathrm {rec}|\) using a linear fit versus \(n_\mathrm {vtx}\), for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates with one or more tracks. The correction varies in the range 90–420 MeV per reconstructed vertex, increasing with \(|\eta |\).

The energy resolution, as determined from simulated data, as a function of the true visible energy after the complete tau calibration is shown in Fig. 16. The resolution is about 20 % at very low E and reduces to about 5 % for energies above a few hundred \(\,\hbox {GeV}\). The resolution is worst in the transition region \(1.3 < {|\eta |}< 1.6\).

Fig. 16
figure 16

Offline energy resolution for hadronically decaying tau leptons, separately for a one and b three associated tracks and for different pseudorapidity regions. The resolution shown is the standard deviation of a Gaussian function fit to the distribution of \((E_\mathrm {reco}-E^\mathrm {true}_\mathrm {vis})/E^\mathrm {true}_\mathrm {vis}\) in a given range of \(E^\mathrm {true}_\mathrm {vis}\) and \(|\eta ^\mathrm {true}_\mathrm {vis}|\)

5.2 Additional offline tau calibration corrections and systematic uncertainties

The systematic uncertainties on the tau energy scale are evaluated with two complementary methods. The deconvolution method gives access to uncertainties on both the absolute TES (differences between reconstructed and true visible energy) and the modelling (differences between data and simulation) and is based on dedicated measurements (such as test beam data and low-luminosity runs) and simulation. The in-situ method only tests the modelling and uses collision data with typical 2012 LHC run conditions. Both methods are also able to provide small additional data-driven corrections albeit only inclusively in \(E_{\mathrm {T}}\) and \(|\eta |\) due to the limited statistical power of the dataset. They thus depend on the first calibration step explained in the previous section to remove any significant TES dependencies on kinematics or pile-up.

The deconvolution method is almost identical to the method employed to measure the jet energy scale for ATLAS in 2010 [49] and is only briefly described here. The central idea is to decompose each tau lepton into its decay products and to combine the calorimeter responses according to the branching ratios of tau leptons to the various hadronic final states. The response to charged hadrons is estimated from different sources depending on the momentum and pseudorapidity; in-situ E / p measurements are used at low momentum, combined test beam measurements are used at high momentum in the central region (\({|\eta |}<0.8\)), and simulation is used otherwise (here, the uncertainty is estimated using events simulated using different hadronic shower models). The response to electromagnetic showers was studied in \(Z \rightarrow ee\) decays and is used for neutral pions. Pseudo-experiments are used to propagate the single-particle response uncertainties to the reconstructed hadronically decaying tau lepton. In each pseudo-experiment, the tau decay product energies are varied randomly using Gaussian distributions centred on the observed ratio of the response in data and simulation and with a width corresponding to the statistical uncertainty, and Gaussian distributions centred at unity and with widths given by each systematic uncertainty. These distributions depend on particle type, energy and pseudorapidity. The TES shift for a single pseudo-experiment is given by the mean of the energy ratio of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)to an identical pseudo-experiment in which only statistical uncertainties of the measurement are considered by Gaussian distributions centred at unity. The distribution of TES shifts for a large number of pseudo-experiments is fitted with a Gaussian function. The mean of the fit is the expected scale shift between data and simulation, and its standard deviation the contribution to the TES uncertainty.

Additional contributions considered are uncertainties due to the detector modelling in the simulation, the underlying event, the effect of pile-up, the non-closure of the calibration method (meaning the difference between the reconstructed and the true \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy, when applying the calibration to the same sample it was derived from) and the hadronic-shower model, as shown in Table 5. The total TES uncertainty for \(E_{\mathrm {T}}>20\) \(\,\hbox {GeV}\) and \({|\eta |}<2.5\) is between 2 and 3 % for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with one track and between 2 and 4 % for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with more tracks, depending on \(E_{\mathrm {T}}\) and \(|\eta |\). A TES shift of 1 % is observed with no significant dependence on \(E_{\mathrm {T}}\) or \(|\eta |\) and a trend towards slightly higher values for 3-track \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates. The shift is dominantly due to E / p response differences between data and simulation.

Table 5 Systematic uncertainties on the tau energy scale estimated using the deconvolution method. In general, the values depend on \(E_{\mathrm {T}}\), \(|\eta |\)  and the number of associated tracks. The range of values for \(E_{\mathrm {T}}>20\) \(\,\hbox {GeV}\) is shown
Fig. 17
figure 17

The \(m_{\mathrm {vis}}\) distribution used for the in-situ offline TES measurement. Shown is the comparison between data and simulation for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with a one or b three associated tracks

The in-situ method is based on the fact that the distribution of the reconstructed visible mass \(m_{\mathrm {vis}}\) in \(Z \rightarrow \tau \tau \) events where one tau decays hadronically and the other to a muon plus neutrinos can be used to measure a TES shift between data and simulation and its uncertainty. Here, \(m_{\mathrm {vis}}\) is defined as the invariant mass of the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)and the muon. The muon momentum scale is measured independently with high precision. The TES shift \(\alpha \) is determined by introducing an energy shift \(E_{\mathrm {T}} \rightarrow (1+\alpha ) E_{\mathrm {T}} \) for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)objects and finding the value \(\alpha \) for which the \(m_{\mathrm {vis}}\) peak position in data and simulation agrees. A fifth-order polynomial fit is used to estimate the \(m_{\mathrm {vis}}\) peak position as simulation studies show that this gives both the highest sensitivity and robustness. For small values of \(\alpha \), the \(m_{\mathrm {vis}}\) peak position depends linearly on \(E_{\mathrm {T}}\).

The results are based on collision data recorded by the ATLAS detector in 2012 using a muon trigger threshold of 24 \(\,\hbox {GeV}\). The event selection is similar to the one used by the \(Z \rightarrow \tau \tau \) tag-and-probe studies described in Sect. 4.1 with the following differences: the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates are required to have \(E_{\mathrm {T}}> 20\) \(\,\hbox {GeV}\) and to satisfy medium tau identification criteria. No selection requirement is applied to \(m_{\mathrm {vis}}\), and a looser \(\cos \Delta \phi >-0.5\) requirement is made. Additionally, a pseudorapidity difference between the \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)and the muon smaller than 1.5 as well as \(E^\tau _{\mathrm {T,vis}}-E_{\mathrm {T}}^\mu > -15\) \(\,\hbox {GeV}\) is required. The motivation for the differences is that this measurement requires a highly pure sample of hadronically decaying tau leptons after applying tau identification while the priority of the efficiency measurement is to obtain a largely unbiased sample before applying any identification requirements.

The background contributions are estimated in the same way as described in Sect. 4.2. The dominant systematic uncertainties of the in-situ measurement are estimated using pseudo-experiments and are due to a potential bias of the fit, missing transverse momentum resolution and scale, muon momentum resolution, muon trigger efficiency and the normalization of the multi-jet background. They are summarized in Table 6.

Table 6 Dominant systematic uncertainties on the tau energy scale estimated using the in-situ method. In general, the values depend on the number of associated tracks. All other systematic uncertainties are smaller than 0.1 %

The measured TES shift is \(\alpha = 0.8\,\% ~\pm ~1.3\) % (stat) \(\pm ~0.6\) % (syst) and \(\alpha =1.1\,\% ~\pm ~1.4\) % (stat) \(\pm ~0.7\) % (syst) for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with one or three associated tracks respectively. No significant dependence on \(\eta \) or pile-up conditions is observed. The corrections are positive, i.e. the momentum of \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)in data has to be scaled up in order to yield agreement (on average) with simulation, and are in agreement with the bias observed in data using the deconvolution method. The resulting \(m_{\mathrm {vis}}\) distribution for data and simulation is shown in Fig. 17 before applying any correction (i.e., \(\alpha =0\)). The uncertainties given above only account for differences between data and simulation and not in the absolute TES. For the latter, uncertainties due to non-closure and pile-up conditions estimated with the deconvolution method have to be added in quadrature to the systematic uncertainties given above.

5.3 Trigger \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy calibration and resolution

As described in Sect. 3.3, reconstructed \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates at both L1 and L2 use a dedicated energy reconstruction algorithm which differs from the offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)energy reconstruction and calibration, while at the EF, the same algorithm is used. In this section, comparisons of the online energy calibrations between data and simulation are shown.

The measured transverse energy resolution for offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates passing medium tau identification is shown in Fig. 18 at all three trigger levels. This measurement is carried out using the same methodology as described in the previous section. The reconstructed energy at L1 is underestimated since at this level calorimeter energies are calibrated at the EM scale. The overestimation seen at L2 is due to the clustering algorithm used at L2, which does not implement the same noise suppression scheme as offline. At the EF, the energy reconstruction is almost identical to the offline case. The slight difference with respect to the offline energy resolution is mainly due to the pile-up corrections, which are only applied offline. Some discrepancies can be seen between the resolutions measured in data and in simulation. This reinforces the importance of having a trigger efficiency measurement performed directly in data as a function of the offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\) \(p_{\mathrm {T}}\), as presented in Sect. 4.2.

Fig. 18
figure 18

The measured tau trigger transverse energy resolution for the offline \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)candidates passing medium tau identification at a L1, b L2 and c the EF. The grey hashed area reflects the statistical uncertainties on the sum of the expected signal and background

6 Summary and conclusions

The algorithms developed in the ATLAS experiment at the LHC for tau identification and tau energy calibration are described, along with their optimization and the associated procedures to mitigate the effects of pile-up. These algorithms were employed in the dataset corresponding to 20.3 fb\(^{-1}\) of \(\sqrt{\mathrm {s}} = 8\) TeV pp collisions. The performance of the tau algorithms have helped to fulfil a variety of physics searches and measurements with hadronically decaying tau leptons, an important part of the ATLAS physics program. The performance of trigger and offline tau identification and calibration is measured, in most cases using \(Z \rightarrow \tau \tau \) tag-and-probe measurements. The uncertainties on the offline tau identification efficiency measurement are dependent on the working point and are about (2–3) % for \({\tau _{\mathrm{had}{\text {-}}\mathrm{vis}}}\)with one associated track, and (4–5) % for the case of three associated tracks, inclusive in \(\eta \) and for a visible transverse momentum greater than 20 \(\,\hbox {GeV}\). A precision of (2–8) % for the tau trigger identification efficiency is measured for hadronic tau decays selected by offline algorithms, depending on the transverse energy. Stability of all algorithms with respect to the pile-up conditions is observed. The reconstructed tau energy scale is measured with a precision of about (2–4) % depending on transverse energy and pseudorapidity, using either a method based on estimating and deconvolving the response uncertainties of the hadronic tau decay products or a direct measurement of the \(Z \rightarrow \tau \tau \) visible mass using collision data recorded in 2012.