Electron reconstruction and identification in the ATLAS experiment using the 2015 and 2016 LHC proton–proton collision

Algorithms used for the reconstruction and identification of electrons in the central region of the ATLAS detector at the Large Hadron Collider (LHC) are presented in this paper; these algorithms are used in ATLAS physics analyses that involve electrons in the final state and which are based on the 2015 and 2016 proton–proton collision data produced by the LHC at √ s = 13 TeV. The performance of the electron reconstruction, identification, isolation, and charge identification algorithms is evaluated in data and in simulated samples using electrons from Z → ee and J/ψ → ee decays. Typical examples of combinations of electron reconstruction, identification, and isolation operating points used in ATLAS physics analyses are shown.


Introduction
Stable particles that interact primarily via the electromagnetic interaction, such as electrons, muons, and photons, are found in many final states of proton-proton ( pp) collisions at the Large Hadron Collider (LHC) located at the CERN Laboratory. These particles are essential ingredients of the ATLAS experiment's Standard Model and Higgs-boson physics programme as well as in searches for physics beyond the Standard Model. Hence, the ability to effectively reconstruct electrons 1 originating from the prompt decay of particles such as the Z boson, to identify them as such with high efficiency, and to isolate them from misidentified hadrons, electrons from photon conversions, and non-isolated electrons originating from heavy-flavour decays are all essential steps to a successful scientific programme.
The ATLAS Collaboration has presented electron-performance results in several publications since the start of the high-energy data-taking in 2010 [1][2][3]. The gradual increase in peak luminosity and the number of overlapping collisions (pile-up) in ATLAS has necessitated an evolution of the electron reconstruction and identification techniques. In addition, the LHC shutdown period of 2013-2014 brought a new charged-particle detection layer to the centre of ATLAS and a restructuring of the trigger system, both of which impact physics analyses with electrons in the final state. These changes require a new benchmarking of electronperformance parameters. The electron efficiency measurements presented in this paper are from the data recorded during the 2015-2016 LHC pp collision run at centre-of-mass energy √ s = 13 TeV. During the period relevant to this paper, the LHC circulated 6.5 TeVproton beams with a 25 ns bunch spacing. The peak delivered instantaneous luminosity was L = 1.37 × 10 34 cm −2 s −1 and the mean number of pp interactions per bunch crossing (hard scattering and pile-up events) was μ = 23.5. The total integrated luminosity [4] used for most of the measurements presented in this paper is 37.1 fb −1 . Another important goal of this paper is to document the methods used by the ATLAS experiment at the start of Run 2 of the LHC (2015 and beyond) to reconstruct, identify, and isolate prompt-electron candidates with high efficiency, as well as to suppress electron-charge misidentification. The methods presented here would be of value to other experiments with similar experimental conditions of fine granularity detection devices but also substantial inactive material in front of the active detector, or with significant activity from pile-up events.
The structure of the paper is described in the following, highlighting additions and new developments with respect to Ref. [3]. Section 2 provides a brief summary of the main components of the detector germane to this paper, with special emphasis on the changes since the 2010-2012 data-taking period. Section 3 itemises the datasets and simulated-event samples used in this paper. Given that the method for calculating efficiencies is common to all measurements, it is described in Sect. 4, before the individual measurements are presented. The algorithms and resulting measurements for electron reconstruction efficiencies are described in Sect. 5, including a detailed discussion of the Gaussian Sum Filter algorithm. Electron identification and the corresponding measurement of efficiencies are described in Sect. 6. New developments here include the optimisation based on simulated events and the treatment of electrons with high transverse momentum. The algorithms used to identify isolated electron candidates and the resulting measured benchmark efficiencies are published for the first time; these are presented in Sect. 7. This paper also presents detailed discussion of studies of the probability to mismeasure the charge of an electron; these are presented in Sect. 8. This section also includes a discussion of the sources of charge misidentification and a new Boosted Decision Tree algorithm that reduces the rate of charge-misidentified electrons significantly. A few examples of combined reconstruction, identification, and isolation efficiencies for typical working points used in ATLAS physics analyses but illustrated with a common Z → ee sample are shown in Sect. 9. The summary of the work is given in Sect. 10.

The ATLAS detector
The ATLAS detector [5] is designed to observe particles produced in the high-energy pp and heavy-ion LHC collisions. It is composed of an inner detector, used for charged-particle tracking, immersed in a 2 T axial magnetic field produced by a thin superconducting solenoid; electromagnetic (EM) and hadronic calorimeters outside the solenoid; and a muon spectrometer. A two-level triggering system reduces the total data-taking rate to approximately 1 kHz. The second level, the high-level trigger (HLT), employs selection algorithms using full-granularity detector information; likelihood-based electron identification and its HLT variant are described in Sect. 6.
The inner detector provides precise reconstruction of tracks within a pseudorapidity range 2 |η| 2.5. The innermost part of the inner detector consists of a high-granularity silicon pixel detector and includes the insertable B-layer [6,7], a new tracking layer closest to the beamline designed to improve impact parameter resolution, which is important primarily for heavy-flavour identification. The silicon pixel detector provides typically four measurement points for charged particles originating in the beam-interaction region. A semiconductor tracker (SCT) consisting of modules with two layers of silicon microstrip sensors surrounds the pixel detector and provides typically eight hits per track at intermediate radii. The outermost region of the inner detector is covered by a transition radiation tracker (TRT) consisting of straw drift tubes filled with a xenon-based gas mixture, interleaved with polypropylene/polyethylene radiators. The TRT offers electron identification capability via the detection of transition-radiation photons generated by the radiators for highly relativistic particles. Some of the TRT modules instead contain an argon-based gas mixture, as mitigation for gas leaks that cannot be repaired without an invasive opening of the inner detector. The presence of this gas mixture is taken into account in the simulation. ATLAS has developed a TRT particle-identification algorithm that partially mitigates the loss in identification power caused by the use of this argon-based gas mixture. For charged particles with transverse momentum p T > 0.5 GeV within its pseudora-pidity coverage (|η| 2), the TRT provides typically 35 hits per track. The ATLAS calorimeter system has both electromagnetic and hadronic components and covers the pseudorapidity range |η| < 4.9, with finer granularity over the region matching the inner detector. The central EM calorimeters are of an accordion-geometry design made from lead/liquidargon (LAr) detectors, providing a full φ coverage. These detectors are divided into two half-barrels (−1.475 < η < 0 and 0 < η < 1.475) and two endcap components (1.375 < |η| < 3.2), with a transition region between the barrel and the endcap (1.37 < |η| < 1.52) which contains a relatively large amount of inactive material. Over the region devoted to precision measurements (|η| < 2.47, excluding the transition regions), the EM calorimeter is segmented into longitudinal (depth) compartments called the first (also known as strips), second, and third layers. The first layer consists of strips finely segmented in η, offering excellent discrimination between photons and π 0 → γ γ decays. At electron or photon energies relevant to this paper, most of the energy is collected in the second layer, which has a lateral granularity of 0.025 × 0.025 in (η, φ) space, while the third layer provides measurements of energy deposited in the tails of the shower. The central EM calorimeter is complemented by two presampler detectors in the region |η| < 1.52 (barrel) and 1.5 < |η| < 1.8 (endcaps), made of a thin LAr layer, providing a sampling for particles that start showering in front of the EM calorimeters. Hadronic calorimetry is provided by the steel/scintillating-tile calorimeter, segmented into three barrel structures within |η| < 1.7, and two copper/LAr hadronic endcap calorimeters. They surround the EM calorimeters and provide additional discrimination through further energy measurements of possible EM shower tails as well as rejection of events with activity of hadronic origin.

Datasets and simulated-event samples
All data collected by the ATLAS detector undergo careful scrutiny to ensure the quality of the recorded information; data used for the efficiency measurements are filtered by requiring that all detector subsystems needed in the analysis (calorimeters and tracking detectors) are operating nominally. After all data-quality requirements (94% efficient), 37.1 fb −1 of pp collision data from the 2015-2016 dataset are available for analysis. Some results in this paper are based on the 2016 dataset only, and contain approximately 10% less data.
Samples of simulated Z → ee and J/ψ → ee decays as well as single-electron samples are used to benchmark the expected electron efficiencies and to define the electronidentification criteria. The Z → ee Monte Carlo (MC) samples were generated with the Powheg-Box v2 MC pro-gram [8][9][10][11][12] interfaced to the Pythia v.8.186 [13] parton shower model. The CT10 parton distribution function (PDF) set [14] was used in the event generation with the matrix element, and the AZNLO [15] set of generator-parameter values (tune) with the CTEQ6L1 [16] PDF set were used for the modelling of non-perturbative effects. The J/ψ → ee samples were generated with Pythia v.8.186; the A14 set of tuned parameters [17] was used together with the CTEQ6L1 PDF set for event generation and the parton shower. The simulated single-electron samples were produced with a flat distribution in η as well as in p T in the region 3.5 GeV to 100 GeV, followed by a linear ramp down to 300 GeV, and then a flat distribution again to 3 TeV. For studies of electrons in simulated event samples, the reconstructed-electron track is required to have hits in the inner detector which originate from the true electron during simulation.
Backgrounds that may mimic the signature of prompt electrons were simulated with two-to-two processes in the Pythia v.8.186 event generator using the A14 set of tuned parameters and the NNPDF2.3LO PDF set [18]. These processes include multijet production, qg → qγ , qq → gγ , W -and Z -boson production (as well as other electroweak processes), and top-quark production. A filter was applied to the simulation to enrich the final sample in electron backgrounds. This filter retains events in which particles produced in the hard scatter (excluding muons and neutrinos) have a summed energy that exceeds 17 GeV in an area of η × φ = 0.1 × 0.1, which mimics the highly localised energy deposits that are characteristic of electrons. When using this background sample, prompt electrons from Wand Z -boson decays are excluded using generator-level simulation information.
Multiple overlaid pp collisions were simulated with the soft QCD processes of Pythia v.8.186 using the MSTW2008LO PDF [19]. The Monte Carlo events were reweighted so that the μ distribution matches the one observed in the data. All samples were processed with the Geant4-based simulation [20,21] of the ATLAS detector.

Electron-efficiency measurements
Electrons isolated from other particles are important ingredients in Standard Model measurements and in searches for physics beyond the Standard Model. However, the experimentally determined electron spectra must be corrected for the selection efficiencies, such as those related to the trigger, as well as particle isolation, identification, and reconstruction, before absolute measurements can be made. These efficiencies may be estimated directly from data using tagand-probe methods. These methods select, from known resonances such as Z → ee or J/ψ → ee, unbiased samples of electrons (probes) by using strict selection requirements on the second object (tags) produced from the particle's decay. The events are selected on the basis of the electron-positron invariant mass. The efficiency of a given requirement can then be determined by applying it to the probe sample after accounting for residual background contamination.
The total efficiency total may be factorised as a product of different efficiency terms: The efficiency to reconstruct in the electromagnetic calorimeter EM-cluster candidates (localised energy deposits) associated with all produced electrons, EMclus , is given by the number of reconstructed EM calorimeter clusters N cluster divided by the number of produced electrons N all . This efficiency is evaluated entirely from simulation, where the reconstructed cluster is associated to a genuine electron produced at generator level. The reconstruction efficiency, reco , is given by the number of reconstructed electron candidates N reco divided by the number of EM-cluster candidates N cluster . This reconstruction efficiency, as well as the efficiency to reconstruct electromagnetic clusters, is described in Sect. 5. The identification efficiency, id , is given by the number of identified and reconstructed electron candidates N id divided by N reco , and is described in Sect. 6. The isolation efficiency is calculated as the number of identified electron candidates satisfying the isolation, identification, and reconstruction requirements N iso divided by N id , and is explained in Sect. 7. Finally, the trigger efficiency is calculated as the number of triggered (and isolated, identified, reconstructed) electron candidates N trig divided by N iso (see for example Ref. [22]; trigger efficiency is not discussed further in this paper). Isolated electrons selected for physics analyses are subject to large backgrounds from misidentified hadrons, electrons from photon conversions, and non-isolated electrons originating from heavy-flavour decays. The biggest challenge in the efficiency measurements presented in this paper is the estimation of probes that originate from background rather than signal processes. This background is largest for the sample of cluster probes, but the fraction of such events is reduced with each efficiency step, from left to right, as given in Eq. (1).
The accuracy with which the detector simulation models the observed electron efficiency plays an important role when using simulation to predict physics processes, for example the signal or background of a measurement. In order to achieve reliable results, the simulated events need to be corrected to reproduce as closely as possible the efficiencies measured in data. This is achieved by applying a multiplicative correction factor to the event weight in simulation. This correction factor is defined as the ratio of the efficiency measured in data to that determined from Monte Carlo events. These correction weights are normally close to unity; deviations from unity usually arise from mismodelling in the simulation of tracking properties or shower shapes in the calorimeters.
Systematic uncertainties in the correction factors are evaluated by varying the requirements on the selection of both the tag and the probe electron candidates as well as varying the details of the background-subtraction method. The central value of the measurement is extracted by averaging the measurement results over all variations. The statistical uncertainty in a single variation of the measurement is calculated following the approach in Ref. [23], i.e. assuming a binomial distribution. If the evaluation of the number of events (before or after the selection under investigation) is the result of a background subtraction, the corresponding statistical uncertainties are also included in the overall statistical uncertainty, rather than in the systematic uncertainty. The systematic uncertainty in the averaged result is obtained from the root-mean-square (RMS) of the individual results, and in the case of non-Gaussian behaviour, it is inflated to cover 68% of the variations.
The tag-and-probe measurements are based on samples of Z → ee and J/ψ → ee events. Whereas the Z → ee sample is used to extract all terms in Eq. (1), the J/ψ → ee sample is only used to extract the identification efficiency id since the significant background as well as the difficulties in designing a trigger for this process prevent its use in determining the reconstruction efficiency. The combination of the two samples allows identification efficiency measurements over a significant transverse energy E T range of 4.5 GeV to 20 GeV for the J/ψ → ee sample, and above 15 GeV (4.5 GeV for the isolation efficiency measurement) for the Z → ee sample, while still providing overlapping measurements between the samples in the E T range 15-20 GeV where the correction factors of the two results are combined using a χ 2 minimisation [2,24]. Combining the correction factors instead of the individual measured and simulated efficiencies reduces the dependence on kinematic differences of the physics processes as they cancel out in the ratio.
Due to the number of events available in the sample, the Z → ee tag-and-probe measurements provide limited information about electron efficiencies beyond approximately electron E T = 150 GeV. The following procedure is used to assign correction factors for candidate electrons with high E T : • reconstruction: the same η-dependent correction factors are used for all E T > 80 GeV, • identification: correction factors determined up to E T = 250 GeV are applicable beyond, • isolation: correction factors of unity are used for E T > 150 GeV.
The following subsections give a brief overview of the methods used to extract efficiencies in the data. Efficiency extraction using simulated events is performed in a very similar fashion, except that no background subtraction is performed. More detailed descriptions may be found in Ref. [3]. 4.1 Measurements using Z → ee events Z → ee events with two electron candidates in the central region of the detector, |η| < 2.47, were collected using two triggers designed to identify at least one electron in the event. One trigger has a minimum E T threshold of 24 GeV (which was changed to 26 GeV during 2016 data-taking), and requires Tight trigger identification (see Sect. 6) and track isolation (see Sect. 7), while the other trigger has a minimum E T threshold of 60 GeV and Medium trigger identification. The tag electron is required to have E T > 27 GeV and to lie outside of the calorimeter transition region, 1.37 < |η| < 1.52. It must be associated with the object that fired the trigger, and must also pass Tight-identification (see Sect. 6) and isolation requirements. If both electrons pass the tag requirements, the event will provide two probes. The invariantmass distribution constructed from the tag electron and the cluster probe is used to discriminate prompt electrons from background. The signal efficiency is extracted in a window of ± 15 GeV around the Z -boson mass peak at 91.2 GeV. Approximately 35 million electron-candidate probes from Z → ee data events are available for analysis.
The probe electrons in the denominator of the reconstruction-efficiency measurement (see Eq. (1)) are electromagnetic clusters both with and without associated tracks, while those in the numerator consist of clusters with matched tracks, i.e. reconstructed electrons (see Sect. 5). These tracks are required to have at least seven hits in the silicon detectors (i.e. both pixel and SCT) and at least one hit in the pixel detector. The background for electron candidates without a matched track is estimated by fitting a polynomial to the sideband regions of the invariant-mass distribution of the candidate electron pairs, after subtracting the remaining signal contamination using simulation. The background for electron candidates with a matched track is estimated by constructing a background template by inverting identification or isolation criteria for the probe electron candidate and normalising it to the invariant-mass sideband regions, after subtraction of the signal events in both the template and the sidebands.
The probe electrons used in the denominator of the identification efficiency measurement are the same as those used in the numerator of the reconstruction efficiency measurement, with an additional opposite-charge requirement on the tag-probe pair; this method assumes that the charge of the candidate is correctly identified. The numerator of the identification measurement consists of probes satisfying the identification criteria under evaluation. Two methods are used in the identification measurements to estimate the nonprompt background [2,3]; they are treated as variations of the same measurement: the Zmass method uses the invariant mass of the tag-probe pair while the Ziso method uses the isolation distribution of probes in the signal mass window around the Z -boson peak. In both cases, and as discussed for the reconstruction-efficiency measurement, background templates are formed and normalised to the sideband regions, after subtraction of the signal events. The contamination from charge-misidentified candidates is negligible in this sample. In the Zmass method, the numerator of the identification efficiency uses same-charge events to obtain a normalisation factor for the template in opposite-charge events, in order to reduce the contamination from signal events.
The isolation-efficiency measurements are performed using the Zmass method, as described above. The denominator in the efficiency ratio is the number of identified electron candidates, while the numerator consists of candidates that also satisfy the isolation criteria under evaluation.
In all cases, systematic uncertainties in the data-MC correction factors are evaluated from the background-subtraction method as well as variations of the quality of the probed electrons via changes in the window around the Z -boson mass peak. They are also evaluated by varying the identification and isolation requirements on the tag, the sideband regions used in the fits, and the template definitions. 4.2 Measurements using J/ψ → ee events J/ψ → ee events with at least two electron candidates with E T > 4.5 GeV and |η| < 2.47 were collected with dedicated dielectron triggers with electron E T thresholds ranging from 4 to 14 GeV. Each of these triggers requires Tight trigger identification and E T above a certain threshold for one trigger object, while only demanding the electromagnetic cluster E T to be higher than some other (lower) threshold for the second object. The J/ψ → ee selection consists of one electron candidate passing a Tight-identification selection (see Sect. 6) and one reconstructed-electron candidate (see Sect. 5). The tag electron is required to be outside the calorimeter transition region 1.37 < |η| < 1.52 and to be associated with the Tight trigger object. The probe electron must be matched to the second trigger object. Due to the nature of the sample (a mixture of prompt and non-prompt decays) as well as significant background, isolation requirements are applied on both the tag and the probe electrons, although for the latter the requirement is very loose so as to not bias the identification-efficiency measurement. Fur-thermore, the tag and the probe electron candidates must be separated from each other in η-φ space by R > 0.15. If both electrons pass the tag requirements, the event will provide two probes. Approximately 80 thousand electroncandidate probes from J/ψ → ee data events are available for analysis.
The invariant-mass distribution of the two electron candidates in the range 1.8-4.6 GeV is fit with functions to extract three contributions: J/ψ events, ψ(2S) events, and the background from hadronic jets, heavy flavour, and electrons from conversions. The J/ψ and ψ(2S) contributions are each modelled with a Crystal Ball function convolved with a Gaussian function, and the background is estimated using same-charge events and fit with a second-order Chebyshev polynomial.
J/ψ → ee events come from a mixture of prompt and non-prompt J/ψ production, with relative fractions depending both on the triggers used to collect the data and on the E T of the probe electrons. Prompt J/ψ mesons are produced directly in pp collisions and in radiative decays of directly produced heavier charmonium states. Non-prompt J/ψ production occurs when the J/ψ is produced in the decay of a b-hadron. Only the prompt production yields isolated electrons, which are expected to have efficiencies similar to those of electrons from physics processes of interest such as H → Z Z * → 4 . Given the difficulties associated with the fact that electrons from nonprompt decays are often surrounded by hadronic activity, two methods have been developed to measure the efficiency for isolated electrons at low E T , both exploiting the pseudoproper time variable 3 t 0 . In the cut method, a requirement is imposed on the pseudo-proper time, so that the prompt component is enhanced, thereby limiting the nonprompt contribution. The residual non-prompt fraction is estimated using simulated samples and ATLAS measurements of J/ψ → μμ [26]. In the fit method, a fit to the pseudo-proper time distribution is used to extract the prompt fraction, after subtracting the background using the pseudoproper time distribution in sideband regions around the J/ψ peak.
The systematic uncertainties in the data-to-simulation correction factors of both methods are estimated by varying the isolation criteria for the tag and the probe electron candidates, the fit models for the signal and background, the signal invariant-mass range, the pseudo-proper time requirement in the cut method, and the fit range in the fit method.

Electron reconstruction
An electron can lose a significant amount of its energy due to bremsstrahlung when interacting with the material it traverses. The radiated photon may convert into an electronpositron pair which itself can interact with the detector material. These positrons, electrons, and photons are usually emitted in a very collimated fashion and are normally reconstructed as part of the same electromagnetic cluster. These interactions can occur inside the inner-detector volume or even in the beam pipe, generating multiple tracks in the inner detector, or can instead occur downstream of the inner detector, only impacting the shower in the calorimeter. As a result, it is possible to produce and match multiple tracks to the same electromagnetic cluster, all originating from the same primary electron.
The reconstruction of electron candidates within the kinematic region encompassed by the high-granularity electromagnetic calorimeter and the inner detector is based on three fundamental components characterising the signature of electrons: localised clusters of energy deposits found within the electromagnetic calorimeter, charged-particle tracks identified in the inner detector, and close matching in η × φ space of the tracks to the clusters to form the final electron candidates. Therefore, electron reconstruction in the precision region of the ATLAS detector (|η| < 2.47) proceeds along those steps, described below in this order. Figure 1 provides a schematic illustration of the elements that enter into the reconstruction and identification (see Sect. 6) of an electron.

Seed-cluster reconstruction
The η × φ space of the EM calorimeter is divided into a grid of 200 × 256 elements (towers) of size η × φ = 0.025 × 0.025, corresponding to the granularity of the second layer of the EM calorimeter. For each element, the energy (approximately calibrated at the EM scale), collected in the first, second, and third calorimeter layers as well as in the presampler (only for |η| < 1.8, the region where the presampler is located) is summed to form the energy of the tower. Electromagnetic-energy cluster candidates are then seeded from localised energy deposits using a sliding-window algorithm [27] of size 3 × 5 towers in η × φ, whose summed transverse energy exceeds 2.5 GeV. The centre of the 3 × 5 seed cluster moves in steps of 0.025 in either the η or φ direction, searching for localised energy deposits; the seedcluster reconstruction process is repeated until this has been performed for every element in the calorimeter. If two seedcluster candidates are found in close proximity (if their towers overlap within an area of η × φ = 5 × 9 units of 0.025 × 0.025), the candidate with the higher transverse energy is retained, if its E T is at least 10% higher than the other candidate. If their E T values are within 10% of each other, the candidate containing the highest-E T central tower is kept. The duplicate cluster is thereby removed. The reconstruction efficiency of this seed-cluster algorithm (effectively EMclus in Eq. (1)) depends on |η| and E T . As a function of E T , it ranges from 65% at E T = 4.5 GeV, to 96% at E T = 7 GeV, to more than 99% above E T = 15 GeV, as can be seen in Fig. 2. This efficiency is determined entirely from simulation. Efficiency losses due to seed-cluster reconstruction for E T > 15 GeV are negligible compared with the uncertainties attributed to the next two steps of the reconstruction (track reconstruction and track-cluster matching).

Track reconstruction
The basic building block for track reconstruction is a 'hit' in one of the inner-detector tracking layers. Charged-particle reconstruction in the pixel and SCT detectors begins by assembling clusters from these hits [28]. From these clusters, three-dimensional measurements referred to as spacepoints are created. In the pixel detector, each cluster equates to one space-point, while in the SCT, clusters from both stereo views of a strip layer must be combined to obtain a threedimensional measurement. Track seeds are formed from sets of three space-points in the silicon-detector layers. The track reconstruction then proceeds in three steps: pattern recog-nition, ambiguity resolution, and TRT extension (for more details of the TRT extension, see Ref. [29]). The patternrecognition algorithm uses the pion hypothesis for the model of energy loss from interactions of the particle with the detector material. However, if a track seed with p T > 1 GeV cannot be successfully extended to a full track of at least seven silicon hits per candidate track and the EM cluster satisfies requirements on the shower width and depth, a second attempt with modified pattern recognition, one which allows up to 30% energy loss for bremsstrahlung at each intersection of the track with the detector material, is made. Track candidates with p T > 400 MeV are fit, according to the hypothesis used in the pattern recognition, using the ATLAS Global χ 2 Track Fitter [30]. Any ambiguity resulting from track candidates sharing hits is resolved at the same stage. In order to avoid inefficiencies for electron tracks with significant bremsstrahlung, if the fit fails under the pion hypothesis and its polar and azimuthal separation to the EM cluster is below a value, a second fit is attempted under an electron hypothesis (an extra degree of freedom, in the form of an additional Gaussian term, is added to the χ 2 to compensate for the additional bremsstrahlung losses coming from electrons; such an energy-loss term is neglected in the pion-hypothesis fit). Figure 2 (top) shows that the reconstruction efficiency of the track-fitting step ranges from 80% at E T = 1 GeV to more than 98% above E T = 10 GeV.  Fig. 2 Top: the total reconstruction efficiency for simulated electrons in a single-electron sample is shown as a function of the true (generator) transverse energy E T for each step of the electron-candidate formation: η × φ = 3 × 5 (in units of 0.025 × 0.025) seed-cluster reconstruction (red triangles), seed-track reconstruction using the Global χ 2 Track Fitter (blue open circles), both of these steps together but instead using GSF tracking (yellow squares), and the final reconstructed electron candidate, which includes the track-to-cluster matching (black closed circles). As the cluster reconstruction requires uncalibrated cluster seeds with E T > 2.5 GeV, the total reconstruction efficiency is less than 60% below 4.5 GeV (dashed line). Bottom: the reconstruction efficiency relative to reconstructed clusters, reco , as a function of electron transverse energy E T for Z → ee events, comparing data (closed circles) with simulation (open circles). The inner uncertainties are statistical while the total uncertainties include both the statistical and systematic components A subsequent fitting procedure, using an optimised Gaussian-sum filter (GSF) [31] designed to better account for energy loss of charged particles in material, is applied to the clusters of raw measurements. This procedure is used for tracks which have at least four silicon hits and that are loosely matched to EM clusters. The separation of the clusterbarycentre position and the position of the track extrapolated from the perigee to the second layer of the calorimeter must satisfy |η cluster − η track | < 0.05 and one of two alterna-tive requirements on the azimuthal separation between the cluster position and the track: −0.20 < φ < 0.05 or −0.10 < φ res < 0.05, where q is the sign of the electric charge of the particle, and φ and φ res are calculated as −q × (φ cluster − φ track ) with the momentum of the track rescaled to the energy of the cluster for φ res . The asymmetric condition for the matching in φ mitigates the effects of energy loss due to bremsstrahlung where tracks with negative (positive) electric charge bend due to the magnetic field in the positive (negative) φ direction.
The GSF method [32] is based on a generalisation of the Kalman filter [33] and takes into account the non-linear effects related to bremsstrahlung. Within the GSF, experimental noise is modelled by a sum of Gaussian functions. The GSF therefore consists of a number of Kalman filters running in parallel, the result of which is that each track parameter is approximated by a weighted sum of Gaussian functions. Six Gaussian functions are used to describe the material-induced energy losses and up to twelve to describe the track parameters. In the final step, the mode of the energy distribution is used to represent the energy loss.
Radiative losses of energy lead to a decrease in momentum, resulting in increased curvature of the electron's trajectory in the magnetic field. When accounting for such losses via the GSF method, all track parameters relevant to the bending-plane are expected to improve. Such a parameter is the transverse impact parameter significance: d 0 divided by its estimated uncertainty σ (d 0 ). Since the curvature, in the ATLAS coordinate frame, is positive for negative particles and negative for positive particles, the signed impact parameter significance (i.e. multiplied by the sign of the reconstructed electric charge q of the electron) is used. Figure 3 shows q × d 0 /σ (d 0 ) for the track associated with the electron, i.e. the primary electron track. A clear improvement in q × d 0 /σ (d 0 ) for genuine electron tracks fitted with the GSF over tracks with the ATLAS Global χ 2 Track Fitter is observed; the distribution is narrower and better centred at zero. Figure 3 also shows, for the ratio of the electroncandidate charge to its momentum q/ p, the relative difference between the true generator value and the reconstructed value; the GSF method shows a sharper and better-centred distribution near zero with smaller tails. The reconstruction efficiency for finding both a seed cluster and a GSF track is shown in Fig. 2 (top).

Electron-candidate reconstruction
The matching of the GSF-track candidate to the candidate calorimeter seed cluster and the determination of the final cluster size complete the electron-reconstruction procedure. This matching procedure is similar to the loose matching discussed above prior to the GSF step, but with stricter requirements; the track-matching in φ is tightened Distributions of the reconstructed electric charge of the candidate electron multiplied by the transverse impact parameter significance, q ×d 0 /σ (d 0 ) (top) and the relative difference between the reconstructed value of the candidate-electron charge divided by its momentum, q/ p, and the true generator value (bottom). The distributions are shown for tracks fitted with the Global χ 2 Track Fitter (dashed red lines) and for tracks fitted with the GSF (solid blue line). The distributions were obtained from a simulated single-electron sample to −0.10 < φ < 0.05, keeping the original alternative requirement −0.10 < φ res < 0.05 the same. If several tracks fulfil the matching criteria, the track considered to be the primary electron track is selected using an algorithm that takes into account the distance in η and φ between the extrapolated tracks and the cluster barycentres measured in the second layer of the calorimeter, the number of hits in the silicon detectors, and the number of hits in the innermost silicon layer; a candidate with an associated track with at least four hits in the silicon layers and no association with a vertex from a photon conversion [34] is considered as an electron candidate. However, if the primary candidate track can be matched to a secondary vertex and has no pixel hits, then this object is classified as a photon candidate (likely a conversion). A further classification is performed using the candidate electron's E/ p and p T , the presence of a pixel hit, and the secondary-vertex information, to determine unambiguously whether the object is only to be considered as an electron candidate or if it should be ambiguously classified as potentially either a photon candidate or an electron candidate. However, this classification scheme is mainly for the benefit of keeping a high photon-reconstruction efficiency. Since all electron identification operating points described in Sect. 6 require a track with a hit in the innermost silicon layer (or in the next-to-innermost layer if the innermost layer is non-operational), most candidates fall into the 'unambiguous' category after applying an identification criterion.
Finally, reconstructed clusters are formed around the seed clusters using an extended window of size 3 × 7 in the barrel region (|η| < 1.37) or 5×5 in the endcap (1.52 < |η| < 2.47) by simply expanding the cluster size in φ or η, respectively, on either side of the original seed cluster. A method using both elements of the extended-window size is used in the transition region of 1.37 < |η| < 1.52. The energy of the clusters must ultimately be calibrated to correspond to the original electron energy. This detailed calibration is performed using multivariate techniques [35,36] based on data and simulated samples, and only after the step of selecting electron candidates rather than during the reconstruction step, which relies on approximate EM-scale energy clusters. The energy of the final electron candidate is computed from the calibrated energy of the extended-window cluster while the φ and η directions are taken from the corresponding track parameters, measured relative to the beam spot, of the track best matched to the original seed cluster.
Above E T = 15 GeV, the efficiency to reconstruct an electron having a track of good quality (at least one pixel hit and at least seven silicon hits) varies from approximately 97-99%. The simulation has lower efficiency than data in the low E T region (E T < 30 GeV) while the opposite is true for the higher E T region (E T > 30 GeV), as demonstrated in Figs. 2 and 4, which show the reconstruction efficiency as a function of E T and as a function of η in bins of E T , respectively, from Z → ee events. All measurements are binned in two dimensions. The uncertainty in the efficiency in data is typically 1% in the E T = 15−20 GeV bin and reaches the per-mille level at higher E T and the uncertainty in simulation is almost an order of magnitude smaller than for data. The systematic uncertainty dominates at low E T for data, with the estimation of background from clusters with no associated track giving the largest contribution. Below E T = 15 GeV, the reconstruction efficiency is determined solely from the simulation; a 2% (5%) uncertainty is assigned in the barrel (endcap) region.

Electron identification
Prompt electrons entering the central region of the detector (|η| < 2.47) are selected using a likelihood-based (LH) identification. The inputs to the LH include measurements from the tracking system, the calorimeter system, and quantities that combine both tracking and calorimeter information. The various inputs are described in Table 1 and the compo-nents of the quantities described in this table are illustrated schematically in Fig. 1. The LH identification is very similar in method to the electron LH identification used in Run 1 (2010-2012) [3], but there are some important differences.
To prepare for the start of data-taking with a higher centerof-mass energy and different detector conditions it was necessary to construct probability density functions (pdfs) based on simulated events rather than data events, and correct the resulting distributions for any mismodelling. Furthermore, the efficiency was smoothed as a function of E T and the likelihood was adjusted to allow its use for electrons with E T > 300 GeV.

The likelihood identification
The electron LH is based on the products for signal, L S , and for background, L B , of n pdfs, P: where x is the vector of the various quantities specified in Table 1. P S,i (x i ) is the value of the signal pdf for quantity i at value x i and P B,i (x i ) is the corresponding value of the background pdf. The signal is prompt electrons, while the background is the combination of jets that mimic the signature of prompt electrons, electrons from photon conversions in the detector material, and non-prompt electrons from the decay of hadrons containing heavy flavours. Correlations in the quantities selected for the LH are neglected. For each electron candidate, a discriminant d L is formed: the electron LH identification is based on this discriminant. The discriminant d L has a sharp peak at unity (zero) for signal (background); this sharp peak makes it inconvenient to select operating points as it would require extremely fine binning. An inverse sigmoid function is used to transform the distribution of the discriminant of Eq. (3): where the parameter τ is fixed to 15 [37]. As a consequence, the range of values of the transformed discriminant no longer varies between zero and unity. For each operating point, a value of the transformed discriminant is chosen: electron candidates with values of d L larger than this value are considered signal. An example of the distribution of a transformed discriminant is shown in Fig. 5 for prompt electrons from Z -boson decays and for background. This distribution illustrates the effective separation between signal and background encapsulated in this single quantity. There are two advantages to using a LH-based electron identification over a selection-criteria-based (so-called "cutbased") identification. First, a prompt electron may fail the cut-based identification because it does not satisfy the selection criterion for a single quantity. In the LH-based selection, this electron can still satisfy the identification criteria, because the LH combines the information of all of the discriminating quantities. Second, discriminating quantities that have distributions too similar to be used in a cut-based identification without suffering large losses in efficiency may be added to the LH-based identification without penalty. Two examples of quantities that are used in the LH-based identification, but not in cut-based identifications, are R φ and f 1 , which are defined in Table 1. Figure 6 compares the distributions of these two quantities for prompt electrons and background.

The pdfs for the LH-identification
The pdfs for the electron LH are derived from the simulation samples described in Sect. 3. As described below, distinct pdfs are determined for each identification quantity in separate bins of electron-candidate E T and η. The pdfs are created from finely binned histograms of the individual identification quantities. To avoid non-physical fluctuations in the pdfs arising from the limited size of the simulation samples, the histograms are smoothed using an adaptive kernel density estimation (KDE) implemented in the TMVA toolkit [37].
Imperfect detector modelling causes differences between the simulation quantities used to form the LH-identification and the corresponding quantities in data. Some simulation quantities are corrected to account for these differences so that the simulation models the data more accurately and hence the determination of the LH-identification operating points is made using a simulation that reproduces the data as closely as possible. These corrections are determined using simulation and data obtained with the Z → ee tag-and-probe method.
The differences between the data and the simulation typically appear as either a constant offset between the quantities (i.e., a shift of the distributions) or a difference in the width, quantified here as the full-width at half-maximum (FWHM) of the distribution of the quantity. In some cases, both shift and width corrections are applied. The quantities f 1 , f 3 , R η , w η2 and R φ have η-dependent offsets, and the quantities f 1 , f 3 , R had , η 1 and φ res have differences in FWHM.
In the case that the difference is a shift, the value in the simulation is shifted by a fixed (η-dependent) amount to make the distribution in the simulation agree better with the distribution in the data. In the case of a difference in FWHM, the value in the simulation is scaled by a multiplicative factor. The optimal values of the shifts and width-scaling factors are determined by minimising a χ 2 that compares the distributions in the data and the simulation. An example of applying an offset is shown in the top panel of Fig. 7, while an example of applying a width-scaling factor is shown in the bottom panel of Fig. 7.
The pdfs for the E T range of 4.5 GeV to 15 GeV are determined using J/ψ → ee Monte Carlo simulation and the Table 1 Type and description of the quantities used in the electron identification. The columns labelled "Rejects" indicate whether a quantity has significant discrimination power between prompt electrons and light-flavour (LF) jets, photon conversions (γ ), or non-prompt electrons from the semileptonic decay of hadrons containing heavy-flavour (HF) quarks (b-or c-quarks). In the column labelled "Usage," an "LH" indi-cates that the pdf of this quantity is used in forming L S and L B (defined in Eq. (2)) and a "C" indicates that this quantity is used directly as a selection criterion. In the description of the quantities formed using the second layer of the calorimeter, 3×3, 3×5, 3×7, and 7×7 refer to areas of η × φ space in units of 0.025 × 0.025    Fig. 5 The transformed LH-based identification discriminant d L for reconstructed electron candidates with good quality tracks with 30 GeV < E T < 35 GeV and |η| < 0.6. The black histogram is for prompt electrons in a Z → ee simulation sample, and the red (dashedline) histogram is for backgrounds in a generic two-to-two process simulation sample (both simulation samples are described in Sect. 3). The histograms are normalised to unit area pdfs for E T > 15 GeV are determined using Z → ee Monte Carlo simulation.

LH-identification operating points and their corresponding efficiencies
To cover the various required prompt-electron signal efficiencies and corresponding background rejection factors needed by the physics analyses carried out within the ATLAS Collaboration, four fixed values of the LH discriminant are used to define four operating points. These operating points are referred to as VeryLoose, Loose, Medium, and Tight in the text below, and correspond to increasing thresholds for the LH discriminant. The numerical values of the discriminant are determined using the simulation. As shown in more detail later in this section, the efficiencies for identifying a prompt electron with E T = 40 GeV are 93%, 88%, and 80% for the Loose, Medium, and Tight operating points, respectively. The identification is optimised in bins of cluster η (specified in Table 2) and bins of E T (specified in Table 3). The selected bins in cluster η are based on calorimeter geometry, detector acceptances and the variation of the material in the   Table 1 and shown for 20 GeV < E T < 30 GeV and 0.6 < |η| < 0.8, that would be inefficient if used in a cut-based identification, but which, nonetheless, have significant discriminating power against background and, therefore, can be used to improve a LH-based identification. In each figure, the red-dashed distribution is determined from a background simulation sample and the black-line distribution is determined from a Z → ee simulation sample. These distributions are for reconstructed electron candidates before applying any identification. They are smoothed using an adaptive KDE and have been corrected for offsets or differences in widths between the distributions in data and simulation as described in Sect. 6.2 inner detector. The pdfs of the various electron-identification quantities vary with particle energy, which motivates the bins in E T . The rate and composition of the background also varies with η and E T .
To have a relatively smooth variation of electron-identification efficiency with electron E T , the discriminant requirements are varied in finer bins (specified in Table 3) than the pdfs. To avoid large discontinuities in electron-identification efficiency at the bin boundaries in electron E T , the pdf values and discriminant requirements are linearly interpolated between the centres of two adjacent bins in E T .
All of the operating points have fixed requirements on tracking criteria: the Loose, Medium, and Tight operating  The f 3 (top) and R had (bottom) pdf distributions in data and simulation for prompt electrons that satisfy 30 GeV < E T < 40 GeV and 0.80 < |η| < 1.15. The distributions for both simulation and data are obtained using the Z → ee tag-and-probe method. KDE smoothing has been applied to all distributions. The simulation is shown before (shaded histogram) and after (open histogram) applying a constant shift ( f 3 , top) and a width-scaling factor (R had , bottom). Although some |η| bins of f 3 additionally have a width-scaling factor, this particular |η| bin only has a constant shift applied points require at least two hits in the pixel detector and seven hits total in the pixel and silicon-strip detectors combined. For the Medium and Tight operating points, one of these pixel hits must be in the innermost pixel layer (or in the next-to-innermost layer if the innermost layer is nonoperational). This requirement helps to reduce the background from photon conversions. A variation of the Loose operating point-LooseAndBLayer-uses the same threshold for the LH discriminant as the Loose operating point and also adds the requirement of a hit in the innermost pixel layer. The VeryLoose operating point does not include an explicit requirement on the innermost pixel layer and requires only one hit in the pixel detector; the goal of this operating point is to provide relaxed identification requirements for background studies. The pdfs of some of the LH quantities-particularly R had and R η -are affected by additional activity in the calorimeter due to pile-up, making them more background-like. The number of additional inelastic pp collisions in each event is quantified using the number of reconstructed primary vertices n vtx . In each η bin and E T bin, the LH discriminant d L is adjusted to include a linear variation with n vtx . Imposing a constraint of constant prompt-electron efficiency with n vtx leads to an unacceptable increase in backgrounds. Instead, the background efficiency is constrained to remain approximately constant as a function of n vtx , and this constraint results in a small (≤ 5 %) decrease in signal efficiency with n vtx .
The minimum E T of the electron identification was reduced from 7 GeV in Run 1 to 4.5 GeV in Run 2. The use of J/ψ → ee to determine LH pdfs at low E T is also new in Run 2. The push towards lower E T was motivated in part by searches for supersymmetric particles in compressed scenarios. In these scenarios, small differences between the masses of supersymmetric particles can lead to leptons with low transverse momentum.
Special treatment is required for electrons with E T > 80 GeV. The f 3 quantity (defined in Table 1) degrades the capability to distinguish signal from background because high-E T electrons deposit a larger fraction of their energy in the third layer of the EM calorimeter (making them more hadron-like) than low-E T electrons. For this reason and since it is not modelled well in the simulation, the pdf for f 3 is removed from the LH for E T > 80 GeV. Furthermore, changes with increasing prompt-electron E T in the R had and f 1 quantities cause a large decrease in identification efficiency for E T > 300 GeV. Studies during development of the identification algorithm showed that this loss in efficiency was very large for the Tight operating point (the identification efficiency fell from 95% at E T = 300 GeV to 73% for E T = 2000 GeV). To mitigate this loss, for electron candidates with E T > 150 GeV, the LH discriminant threshold for the Tight operating point is set to be the same as for the Medium operating point, and two additional selection criteria are added to the Tight selection: E/ p and w stot . The requirement on w stot depends on the electron candidate η, while the requirement on E/ p is E/ p < 10. The high value of the latter requirement takes into account the decreased momentum resolution in track fits of a few 100 GeV and above. With these modifications, good signal efficiency and background rejection are maintained for very high E T electrons in searches for physics beyond the Standard Model, such as W → eν.
In Run 1, electron candidates satisfying tighter operating points did not necessarily satisfy the more efficient looser operating points. This situation was a result of using different quantities in the electron LH for the different operating points. In Run 2, electron candidates satisfying tighter operating points also satisfy less restrictive operating points, i.e. an electron candidate that satisfies the Tight criteria will also pass the Medium, Loose, and VeryLoose criteria.
Another important difference in the electron identification between Run 1 and Run 2 is that the LH identification is used in the online event selection (the high-level trigger, HLT) in Run 2, instead of a cut-based identification in Run 1. This change helps to reduce losses in efficiency incurred by applying the offline identification criteria in addition to the online criteria. The LH identification in the trigger is designed to be as close as possible to the LH used in offline data analysis; however, there are some important differences. The p/ p quantity is removed from the LH because it relies on the GSF algorithm (see Sect. 5.2), which is too CPU-intensive for use in the HLT. The average number of interactions per bunch crossing, μ , is used to quantify the amount of pileup, again because the determination of the number of primary vertices, n vtx , is too CPU-intensive for the HLT. Both the d 0 and d 0 /σ (d 0 ) quantities are removed from the LH used in the trigger in order to preserve efficiency for electrons from exotic processes which might have non-zero track impact parameters. Finally, the LH identification in the trigger uses quantities reconstructed in the trigger, which generally have poorer resolution than the same quantities reconstructed offline. The online operating points corresponding to VeryLoose, Loose, Medium, and Tight are designed to have efficiencies relative to reconstruction like those of the corresponding offline operating points. Due to these differences, the inefficiency of the online selection for electrons fulfilling the same operating point as the offline selection is typically a few percent (absolute), up to 7% for the Tight operating point.
The efficiencies of the LH-based electron identification for the Loose, Medium, and Tight operating points for data and the corresponding data-to-simulation ratios are summarised in Figs. 8 and 9. They are extracted from J/ψ → ee and Z → ee events, as discussed in Sect. 4. The variations of the efficiencies with E T , η, and the number of reconstructed primary vertices are shown. Requirements on the transverse (d 0 ) and longitudinal (z 0 ) impact parameters measured as the distance of closest approach of the track to the measured primary vertex (taking into account the beam-spot and the tilt of the beam-line) are applied when evaluating the numerator of the identification efficiency. For the Tight operating point, the identification efficiency varies from 55% at E T = 4.5 GeV to 90% at E T = 100 GeV, while it ranges from 85% at 123 E T = 20 GeV to 96% at E T = 100 GeV for the Loose operating point. The uncertainties in these measured efficiencies for the Loose (Tight) operating point range from 3% (4%) at E T = 4.5 GeV to 0.1% (0.3%) for E T = 40 GeV. As mentioned earlier in this section, simulation was used to determine the discriminant values that define the various operating points, with the intended outcome that the efficiencies would fall smoothly with decreasing electron E T , while keeping the rapidly increasing background at acceptable levels. The simple offsets and width variations applied to the simulation to account for mismodelling of the EM-calorimeter shower shapes (see Sect. 6.2) work well at higher electron E T , but are unable to fully correct the simulation at lower electron E T . This leads to an unintended larger efficiency in data for signal electrons at lower E T , as can be seen in Fig. 8. The figure  also shows the corresponding rise in the data-to-simulation ratios.
The lower efficiencies of the Medium and Tight operating points compared to Loose result in an increased rejection of background; the rejection factors for misidentified electrons from multijet production (evaluated with the two-totwo process simulation sample described in Sect. 3) increase typically by factors of approximately 2.5 for Medium and 5 for Tight compared to Loose, in the E T range of 4-50 GeV. Computations and measurements of the rejection, especially absolute rejections, are typically associated with large uncertainties due to ambiguities in the definition of the denominator, and the diversity of the sources of background. The factors mentioned above are similar to those published in Table 3 of the ATLAS Run-1 publication [3] when these considerations are taken into account.

Electron isolation
A considerable challenge at the LHC experiments is to differentiate the prompt production of electrons, muons, and photons in signal processes (from the hard-scattering vertex, or from the decay of heavy resonances such as Higgs, W , and Z bosons) from background processes such as semileptonic decays of heavy quarks, hadrons misidentified as leptons and photons, and photons converting into electron-positron pairs in the detector material upstream of the electromagnetic calorimeter. A characteristic signature of such a signal is represented by little activity (both in the calorimeter and in the inner detector) in an area of η × φ surrounding components. The bottom panel shows the data-to-simulation ratios the candidate object. However, the production of boosted particles decaying, for example, into collimated electronpositron pairs or the production of prompt electrons, muons, and photons within a busy experimental environment such as in tt production can obscure the picture. Variables are constructed that quantify the amount of activity in the vicinity of the candidate object, something usually performed by summing the transverse energies of clusters in the calorimeter or the transverse momenta of tracks in a cone of radius R = ( η) 2 + ( φ) 2 around the direction of the electron candidate, excluding the candidate itself.
Several components enter into building such isolation variables: identifying the candidate object itself, its direction, and its contribution to the activity within the cone, and summing, in a pile-up and underlying-event robust way, the other activity found within the cone. The two classes of isolation variables considered in this paper are based on calorimeter and tracking measurements, and are respectively discussed in Sects. 7.1 and 7.2.

Calorimeter-based isolation
The reconstruction of electron candidates is described in Sect. 5. To build an isolation variable, a cone of size R is then delineated around the candidate electron's cluster position.
The computation of calorimeter-based isolation in the early running period of ATLAS simply summed the transverse energies of the calorimeter cells (from both the electromagnetic and hadronic calorimeters) within a cone aligned with the electron direction, excluding the candidate's contribution. This type of calorimeter-based variable exhibited a lack of pile-up resilience and demonstrated poor data-simulation agreement. A significant improvement was achieved by using the transverse energies of topological clusters [38] instead of cells, thus effectively applying a noisesuppression algorithm to the collection of cells.
Topological clusters are seeded by cells with a deposited electromagnetic-scale energy of more than four times the expected noise-level threshold of that cell; this includes both electronic noise and the effects of pileup. The clusters are then expanded, in the three spatial directions across all electromagnetic and hadronic calorimeter layers, by iteratively adding neighbouring cells that contain a deposited energy more than twice the noise level. After the expansion around the cluster stops due to a lack of cells satisfying the energy threshold requirements, a final shell of cells surrounding the agglomeration is added to the cluster. The topological clusters used in the isolation computation are not further calibrated: they remain at the electromagnetic scale, regardless of the origin of the particle.
The energies of all positive-energy topological clusters, whose barycentres fall within the cone of radius R, as illustrated in Fig. 10, are summed into the raw isolation energy variable E isol T,raw . This raw isolation energy still includes the energy deposited by the candidate electron, called the core energy E T,core . The core energy is subtracted by removing the cells included in a η × φ = 0.125 × 0.175 rectangle around the candidate's direction, as illustrated in Fig. 10. The advantage of this method is its simplicity and stable subtraction scheme for both the signal and background candidates. A disadvantage of this method is that the candidate object may deposit energy outside of this fixed rectangular area which may be incorrectly assigned as additional activity, requiring an additional leakage correction to the subtracted core energy. The core leakage correction is evaluated using samples of simulated single electrons (without additional pile-up activity). The energy leaking into the cone is then fit to a Crystal Ball function; its most probable value μ CB is parameterised as a function of E T and is used as an estimator of the average leakage, E T,leakage (E T ). The corrections are derived in ten bins of the associated cluster η position. Figure 11 shows the isolation energy corrected with a rectangular core, without and with the calculated leakage correction, as a function of the electron E T for a sample of simulated single electrons which includes the effects of pile-up; a 123 φ η Fig. 10 Schema of the calorimeter isolation method: the grid represents the second-layer calorimeter cells in the η and φ directions. The candidate electron is located in the centre of the purple circle representing the isolation cone. All topological clusters, represented in red, for which the barycentres fall within the isolation cone are included in the computation of the isolation variable. The 5 × 7 cells (which cover an area of η × φ = 0.125 × 0.175) represented by the yellow rectangle correspond to the subtracted cells in the core subtraction method leakage correction is essential when using a rectangular core region.
The pile-up and underlying-event contribution to the isolation cone is estimated from the ambient energy density [39]. For each event, the entire calorimeter acceptance up to |η| = 5 is used to gather positive-energy topological clusters using the k t jet-clustering algorithm [40,41] with radius parameter R = 0.5, with no jet p T threshold. The area A of each jet is estimated and the transverse energy density ρ of each jet is computed as ρ = p T /A. The median energy density ρ median of the distribution of jet densities in the event is used as an estimator of the transverse energy density of the event. For a simulated Z → ee sample at √ s = 13 TeV with average pile-up μ = 22, ρ median is approximately 4 GeV per unit of η − φ space in the central η region of the calorimeter, decreasing to 2 GeV at |η| = 2.5. The pileup/underlying-event correction is then evaluated as: where R is the radius of the isolation cone and A core is the area of the signal core that was subtracted. The η dependence of ρ is estimated in two bins: |η| < 1.5 and 1.5 < |η| < 3.0. Figure 11 shows the size of this pile-up correction for a simulated single-electron sample with μ = 13.5 pile-up. The fully corrected calorimeter-based isolation variable E isol T,cone calculated within the cone of radius R is then obtained after subtracting the components described above, namely Figure 11 shows the resulting distribution as a function of the transverse energy of the electron for a simulated singleelectron sample. The distribution is slightly positive since only positive-energy clusters are summed, allowing for only positive fluctuations from noise.

Track-based isolation
The computation of track-based isolation variables uses tracks with p T > 1 GeV, reconstructed within a fiducial region of the inner detector, |η| < 2.5, and that satisfy basic track-quality requirements. This track selection, optimised using candidate muons from simulated tt samples, includes a minimum number of hits identified in the silicon detectors and a maximum number of inoperable detector regions crossed by the track. In order to minimise the impact of pile-up, a requirement is placed on the longitudinal impact parameter, z 0 , corrected for the reconstructed position of the primary vertex and multiplied by the sine of the track polar angle: |z 0 sin θ | < 3 mm. This requirement on |z 0 sin θ | aims to select tracks that originate from the vertex that is chosen to be the relevant vertex of the process. In most cases, the rele-vant vertex corresponds to the "hardest" vertex of the event, i.e. the vertex for which the sum of the squares of the transverse momenta of the associated tracks is the largest; this is the vertex used by default in the track-isolation computation. Track-based isolation variables are constructed by summing the transverse momenta of the tracks found within a cone of radius R aligned with the electron track, excluding the candidate's own contribution.
The trackp T contribution of the candidate electron to the track-isolation variable must be subtracted from the cone. Electrons can undergo bremsstrahlung radiation with the radiated photons converting into secondary electrons; such additional particles should be counted as part of the initial particle's energy. For this reason, tracks are extrapolated to the second layer of the EM calorimeter. All extrapolated tracks that fall within a η × φ = 0.05 × 0.1 window around the cluster position are considered to be part of the candidate and are removed from the track-isolation-variable computation. The resulting track-isolation variable is called p isol T . Unlike calorimeter isolation, where a cone with a radius much less than R = 0.2 is difficult to build due to the finite granularity of the calorimeter, the much smaller tracker granularity allows the use of narrower cone sizes. For example, in boosted decay signatures or very busy environments, other objects can be close to the signal lepton direction. For such cases, a variable-cone-size track isolation, p isol T,var , can be used, one that progressively decreases in size as a function of the p T of the candidate: where R max is the maximum cone size (typically 0.2-0.4). The value of 10 GeV in the argument is derived with a simulated tt sample, and designed to maximise the rejection of background.

Optimisation of isolation criteria and resulting efficiency measurements
The implementation of isolation criteria is specific to the physics analysis needs, be it to identify isolated prompt electrons or electrons produced in a busy environment, or to reject light hadrons misidentified as electrons. Precision measurements with copious signal at lower p T may favour tighter isolation requirements, and be willing to sacrifice some signal in order to ensure high background rejection, whereas searches at high p T may instead favour looser requirements in order to maintain high signal efficiency. Therefore, several isolation operating points were established that use calorimeter-based isolation in a cone of radius R = 0.2 (Sect. 7.1) or trackbased isolation using a variable-size cone with R max = 0.2 or 0.4 (Sect. 7.2), or both types of isolation simultaneously. The requirements for each efficiency-targeted operating point are established in bins of electron E T and η with edges: The operating points are defined in three categories: • targeting a fixed value of the isolation efficiency iso , uniform in the E T , η of the electron ('Loose' isolation), • targeting a fixed value of the isolation efficiency iso , dependent on the E T of the electron but uniform in η ('Gradient' isolation), • imposing fixed requirements on the value of the isolation variable ('Fix'). Figure 12 shows the isolation efficiencies measured in data and the corresponding data-to-simulation ratios as a function of the electron E T and η for the operating points given in Table 4 and for candidate electrons satisfying Tight identification requirements. The efficiencies that determine the values of the requirements given in Table 4 are evaluated in simulation from a J/ψ → ee sample for E T < 15 GeV and from a Z → ee sample for E T > 15 GeV. Since the value of the efficiency is process dependent, this can result in a jump in efficiency at the transition point of E T = 15 GeV, as can be seen in the top-left plot of Fig. 12, which was produced with a Z → ee sample over the full E T range. The overall differences between data and simulation are less than approximately 2-4%; the operating point 'Fix (Tight)' has the most significant difference in shape in η.

Reconstruction of electric charge
The electric charge of an electron is determined from the curvature of the associated track reconstructed in the inner detector. The misidentification of electron charge can result from the matching of an incorrect track to the electron candidate or from a mismeasurement of the curvature of the primary electron track.
The probability of bremsstrahlung emission and subsequent photon conversion to an electron-positron pair (see Sect. 5) depends significantly on the amount of detector material traversed. As shown in Fig. 13 (top, left) for a simulated Z → ee sample, most electrons for which the wrong charge was assigned are found in the pseudorapidity region 1.04

Fig. 12
Isolation efficiencies for data (in the upper panels) and the ratio to simulation (lower panels) for the operating points given in Table 4 as a function of candidate-electron E T (left) and η (right) in Z → ee events. Top plots: efficiencies for optimised operating points targeting specific efficiencies. Bottom plots: efficiencies for operating points where fixed requirements are applied to the isolation variables. Cone sizes of R = 0.2 for calorimeter isolation and R max = 0.2 for track isolation are used (except for "Fix (Track R max = 0.4)" where a twice larger radius cone is used). The inner uncertainties are statistical while the total uncertainties include both the statistical and systematic components 1.5 < |η| < 2.2, corresponding to a region in the detector with a relatively large amount of inactive material. Pair production resulting from bremsstrahlung is the main reason why a wrong track can be matched to the electron candidate; three tracks in close proximity are present, two of which have the correct charge assignment, causing an ambiguity in the selection of the primary electron track. In this context, "correct track" means the primary electron track while "correct charge" means the track has the same reconstructed charge as the primary electron. The number of reconstructed tracks passing the requirements described in Sect. 5.3 is shown in Fig. 13 (top, right); events with electrons   incorrectly reconstructed charge with respect to the particle that caused the track and the correctly and incorrectly assigned track with respect to the primary electron. The distributions are normalised to the same area that have a wrongly assigned track tend to have higher track multiplicities. Electrons produced in conversions of photons originating from the primary electron have a most probable value of three tracks regardless of whether the charge was correctly reconstructed or not, illustrating the probabilistic nature of the process.
The presence of additional electrons and positrons in close proximity can cause the track to be reconstructed from hits of primary and secondary electrons and positrons. Significant energy loss can also distort the shape of the track. At very high energies or large pseudorapidities, the tracks can also become straighter and hence it is difficult to determine their curvature. All of these effects impact the ability to correctly fit the track, as can be seen from the comparison of the correct charge and wrong charge distributions for the correctly assigned track. These effects are more pronounced as a function of increased passive material, located at large values of η, as shown by comparing the equivalent two distributions in Fig. 13 (top, left). Figure 13 (bottom, left) and (bottom, right) show the distributions of the electric charge multiplied by the transverse impact parameter significance q × d 0 /σ (d 0 ) and the E/ p, respectively, for electron candidates with correctly and incorrectly reconstructed charge and assigned track. As expected, electrons with incorrectly reconstructed charge tend to have a larger magnitude of d 0 /σ (d 0 ). Their q × d 0 /σ (d 0 ) tends to be more negative due to the incorrect measurement of the electric charge. The distributions have a significant fraction of events with E/ p = 1, due to either the wrong track being matched to the candidate electron or the wrong track curvature being measured, resulting in an erroneous measurement of the electron's momentum.

Suppression of charge misidentification
The charge-misidentification rate for reconstructed electron candidates can be reduced with an additional selection criterion based on the output discriminant of a boosted decision tree (BDT) trained using the TMVA toolkit. The training of the BDT is performed on a simulated single-electron sample. A set of eight variables related to the quality of the chosen primary electron track, the lateral development of the electromagnetic shower (which is expected to be narrower for electrons that don't undergo bremsstrahlung), and combined tracking and calorimeter quantities is used to separate electrons with the correctly reconstructed electric charge from those reconstructed with the wrong charge. The list of these quantities, ranked by their contribution to the discriminating power of the selection (highest to lowest, as evaluated with a simulated Z → ee sample), is: E/ p, φ res , q ×d 0 , the pseudorapidity η of the electron, R φ , (q/ p)/σ (q/ p), φ 1 , and the transverse energy E T of the electron. Table 1 provides the definitions of these variables except for the charge-to-momentum ratio q/ p divided by its uncertainty σ (q/ p) and the distance in φ between the cluster position in the first layer of the calorimeter and the extrapolated track, φ 1 .
The BDT parameters are optimised to achieve the best possible rejection of electrons reconstructed with the wrong charge given an efficiency loss of 3% for electrons with correctly reconstructed charge. The resulting efficiencies from applying the additional BDT requirement for electrons already identified with a Medium requirement (see Sect. 6) are shown in Fig. 14 for candidate electrons from Z → ee events. The inclusion of the kinematic properties E T and η in the BDT introduces a dependence of the efficiency on these quantities (in particular for η), even for electron candidates reconstructed with the correct charge. The differences observed between data and simulation arise not only from mismodelling of some of the electromagnetic shower shapes in the simulation but also from the modelling of complicated objects such as the beam-spot. For this reason, data-tosimulation scale factors are calculated; they are also shown in Fig. 14. At very high |η|, the efficiency in data is approximately 76%. The corresponding efficiency in simulation is about 8% higher, due to mismodelling. Analyses using samecharge dilepton pairs such as searches for supersymmetry in same-charge signatures [42] or the search for Higgs boson production in association with a top-quark pair [43] benefit from a suppression of electrons from charge misidentification with the BDT requirement.

Measurement of the probability of charge misidentification
A measurement of the probability of charge misidentification is performed using a sample of Z → ee events where both electrons are required to satisfy the same identification and isolation criteria. The probability for an electron to be reconstructed with the wrong charge in a given (η, E T ) bin is obtained from the numbers of events where both electrons have the same (SC) and opposite electric charges (OC). The average expected number of same-charge events for a pair of electrons falling into their respective (η, E T ) bins i and j, N SC i j , follows from the total number of observed events n OC+SC i j in these same two (η, E T ) bins regardless of charge, corrected for the charge-misidentification probability in each bin i, j : The probabilities i and j are obtained by maximising the Poisson likelihood to observe the number of same-charge events n SC i j in these bins. The non-prompt-background contributions to n OC i j and n SC i j are estimated by taking equal-size sideband invariant-mass regions about the Z -boson peak and data; closed points: simulation; and lower panels: data-to-simulation ratios. The uncertainties are smaller than the markers and hence not visible linearly interpolating to estimate the number of events under the mass peak. Systematic uncertainties associated with the Z -boson mass window are evaluated by increasing and decreasing its range. Systematic uncertainties from other sources are evaluated by not performing the background subtraction, by taking differences between probabilities in events where both electrons fired the trigger and probabilities in events where only one electron fired the trigger, and by comparing the charge-misidentification rates extracted using the likelihood method with the rates extracted from a comparison with the true charge as determined in simulation. Figure 15 shows the charge-misidentification rates in data and simulation for Medium electrons in Z → ee events as a function of E T and |η|, demonstrating the impact of applying the BDT requirement to suppress charge misidentification, as explained in Sect. 8.2. As expected, the charge-misidentification rates are larger at high E T and |η|. Two identification operating points are also contrasted showing the impact of using a Tight identification requirement without the BDT requirement compared with using Medium identification but also including this BDT requirement.
The knowledge of electron charge-misidentification rates is crucial for the estimation of background from charge misidentification in measurements, e.g. in the observation of the electroweak W ± W ± j j production [44] or the search for doubly charged Higgs boson production [45]. To account for disagreements between data and simulation, data-to-simulation ratios, extracted from the efficiencies described above, are provided as correction factors for analyses to account for this mismodelling in the simulation.

Usage of electron selections in physics measurements
Efficiencies for electron reconstruction, identification, isolation, and charge identification are shown for several operating points in Sects. 5, 6, 7, and 8, respectively. The product of these efficiencies (and their corresponding data-tosimulation correction factors) is relevant in physics measurements. In this section, electron efficiencies for data from a Z → ee sample are presented for the combination reco × id × iso (see Eq. (1)) for some typical operating points used in ATLAS physics analyses. The results are shown in Fig. 16   showing the impact of applying the BDT requirement to suppress charge misidentification (red squares vs. blue triangles). Medium identification with BDT vs. Tight identification without BDT is also explored (black circles vs. blue triangles). The Fix (Tight) isolation requirement is applied in all cases (see Table 4). The inner uncertainties are statistical while the total uncertainties include both the statistical and systematic components  syst. unc. ⊕ stat. Fig. 16 The product of reconstruction, identification, and isolation efficiencies reco × id × iso for data from a Z → ee sample as a function of electron E T (top) and η for E T > 4.5 GeV (bottom), for the following operating points: "Loose + Fix (Loose)" (black circles), "Medium + Gradient (Loose)" (yellow upside-down triangles), "Tight + Gradient" (blue squares) and "Tight + Fix" (Track R max = 0.4) (red triangles). The inner uncertainties are statistical while the total uncertainties include both the statistical and systematic components. The lower panels show data-to-simulation ratios as well as the relative statistical and total uncertainties (statistical and systematic added in quadrature) applicable to both the data efficiencies and correction factors

Conclusions
The focus of this paper is to document the methods used by the ATLAS experiment at the LHC in Run 2 to reconstruct, identify, and isolate prompt-electron candidates with high efficiency, as well as to suppress electron-charge misidentification. Efficiency measurements as evaluated with 37.1 fb −1 of 13 TeV pp collision data recorded in 2015-2016 are then compared with those determined from simulation.
The method used to reconstruct electron candidates is explained for candidates in the central region of the detector (|η| < 2.47). Illustrative measurements of the reconstruction efficiency are presented for candidates with transverse energy E T > 4.5 GeV. For E T > 15 GeV, the reconstruction efficiencies range from 96 to 99%, with the data uncertainties typically 1% for E T = 15−20 GeV and reaching the per-mille level at higher E T .
The likelihood method used to identify electrons given the existence of a reconstructed electron candidate is also explained for candidates in the central region of the detector (|η| < 2.47). The minimum E T of the identified electron is reduced from 7 GeV in Run 1 to 4.5 GeV in Run 2. Benchmark efficiency measurements for three identification operating points are provided: for the Tight operating point, the identification efficiency varies from 55% at E T = 4.5 GeV to 90% at E T = 100 GeV, while it ranges from 85% at E T = 20 GeV to 96% at E T = 100 GeV for the Loose operating point. The uncertainties in these measured efficiencies for the Loose (Tight) operating point range from 3% (4%) at E T = 4.5 GeV to 0.1% (0.3%) for E T = 40 GeV.
The ability to identify electron candidates isolated from any other local activity in the detector is documented. Two varieties of isolation variables, calorimeter-and trackingbased, are considered. Since the implementation of isolation criteria is specific to the physics analysis needs, several operating points are illustrated; their typical isolation efficiencies are measured in data and determined from simulation, ranging from approximately 90% for the tightest operating points to nearly 99% for the loosest, for electrons with E T = 40 GeV.
Possible sources of electron-charge misidentification are explored and an algorithm to suppress such effects is outlined. The efficiency for correct charge reconstruction is illustrated for candidates passing the Medium identification requirements. The efficiency is approximately 96% in both data and simulation and is mostly independent of E T , while it varies from 90 to 99% over most of the electron η region except at very high |η| where the efficiency in data is approximately 76%. The corresponding efficiency in simulation is about 8% higher, due to mismodelling.
Finally, since total efficiencies are used in ATLAS physics analyses, the product of the electron efficiencies reco × id × iso measured in data and presented in this paper for some typ-ical operating points demonstrate the impact of these efficiencies on ATLAS measurements and searches for new physics. CERN

Data Availability Statement
This manuscript has no associated data or the data will not be deposited. [Authors' comment: All ATLAS scientific output is published in journals, and preliminary results are made available in Conference Notes. All are openly available, without restriction on use by external parties beyond copyright law and the standard conditions agreed by CERN. Data associated with journal publications are also made available: tables and data from plots (e.g. cross section values, likelihood profiles, selection efficiencies, cross section limits, ...) are stored in appropriate repositories such as HEPDATA (http:// hepdata.cedar.ac.uk/). ATLAS also strives to make additional material related to the paper available that allows a reinterpretation of the data in the context of new theoretical models. For example, an extended encapsulation of the analysis is often provided for measurements in the framework of RIVET (http://rivet.hepforge.org/)." This information is taken from the ATLAS Data Access Policy, which is a public document that can be downloaded from http://opendata.cern.ch/record/413 [opendata.cern.ch].] Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecomm ons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. Funded by SCOAP 3 .