Measurement of the charge asymmetry in top-quark pair production in the lepton-plus-jets final state in pp collision data at $$\sqrt{s}=8\,\mathrm TeV{}$$s=8TeV with the ATLAS detector

This paper reports inclusive and differential measurements of the $$t\bar{t}$$tt¯ charge asymmetry $$A_{\text {C}}$$AC in $$20.3~{\mathrm{fb}^{-1}}$$20.3fb-1 of $$\sqrt{s} = 8~\mathrm TeV{}$$s=8TeV$$pp$$pp collisions recorded by the ATLAS experiment at the Large Hadron Collider at CERN. Three differential measurements are performed as a function of the invariant mass, transverse momentum and longitudinal boost of the $$t\bar{t}$$tt¯ system. The $$t\bar{t}$$tt¯ pairs are selected in the single-lepton channels (e or $$\mu $$μ) with at least four jets, and a likelihood fit is used to reconstruct the $$t\bar{t}$$tt¯ event kinematics. A Bayesian unfolding procedure is performed to infer the asymmetry at parton level from the observed data distribution. The inclusive $$t\bar{t}$$tt¯ charge asymmetry is measured to be $$A_{\text {C}}{} = 0.009 \pm 0.005$$AC=0.009±0.005 (stat. $$+$$+ syst.). The inclusive and differential measurements are compatible with the values predicted by the Standard Model.


Introduction
The 8 TeV proton-proton ( pp) collision data delivered by the CERN Large Hadron Collider (LHC) represents a unique laboratory for precision measurements of the top-quark properties. One interesting feature of tt production is the difference in rapidity between top quarks and top antiquarks. In pp collisions, this distinct behaviour of top quarks and antiquarks is called the charge asymmetry, A C [defined in Eq. (1)]. The Standard Model (SM) expectation computed at next-to-leading order (NLO) in quantum chromodynamics (QCD), including electroweak corrections, predicts A C to be at the one percent level [1]. Previous asymmetry measurements at the LHC by both the CMS and ATLAS collaborations based on the 7 TeV data, and by the CMS collaboration based on the 8 TeV data, do not report any significant deviation from the SM predictions [2][3][4][5][6][7]. Charge asymmetry measurements are largely limited by the size of the available data sample, and therefore the larger dataset recorded by the ATLAS detector at √ s = 8 TeV allows for an improvement on the precision of the measurement from the √ s = 7 TeV dataset.
At hadron colliders, tt production is predicted to be symmetric under the exchange of top quark and antiquark at leading order (LO). At NLO, the process qq → tt g develops an asymmetry in the top-quark rapidity distributions, due to interference between processes with initial-and final-state gluon emission. The interference between the Born and the NLO diagrams of the qq → tt process also produces an asymmetry. The qg → ttg production process is also asymmetric, but its contribution is much smaller than that from qq.
In qq scattering processes in pp collisions at the Tevatron, the direction of the incoming quark almost always coincides with that of the proton, and this knowledge of the direction of the incoming quarks allows one to define a direct measurement of the forward-backward asymmetry, A FB [8][9][10][11]. In pp collisions at the LHC, since the colliding beams are symmetric, it is not possible to use the direction of the incoming quark to define an asymmetry. However, valence quarks carry on average a larger fraction of the proton momentum than sea antiquarks, hence top quarks are more forward and top antiquarks are more central. Using this feature it is possible to define a forward-central asymmetry for the tt production, referred to as the charge asymmetry, A C [8, 12,13]: where |y| ≡ |y t | − |yt | is the difference between the absolute value of the top-quark rapidity |y t | and the absolute value of the top-antiquark rapidity |yt |. At the LHC, the dominant mechanism for tt production is the gluon fusion process, while production via the qq or the qg interactions is small. Since gg → tt processes are charge-symmetric, they only contribute to the denominator of Eq. (1), thus diluting the asymmetry. Several processes beyond the Standard Model (BSM) can alter A C [12,[14][15][16][17][18][19][20][21][22][23][24][25], either with anomalous vector or axialvector couplings (e.g. axigluons) or via interference with SM processes. Different models also predict different asymmetries as a function of the invariant mass m tt , the transverse momentum p T,tt and the longitudinal boost β z,tt along the z-axis 1 of the tt system [26]. The interest in precisely measuring charge asymmetries in top-quark pair production at the LHC has grown after the CDF and D0 collaborations reported measurements of A FB that were significantly larger than the SM predictions, in both the inclusive and differential case as a function of m tt and of the rapidity of the tt system, y tt [10,11,[27][28][29][30]. For the most general BSM scenarios [31], the A C measurements from the LHC are still compatible with the Tevatron results. However, for specific simple models [20], tension still exists between the LHC and Tevatron results. This motivates the interest in a more precise measurement of the tt production charge asymmetry at the LHC.
In this paper, a measurement of the tt production charge asymmetry in the single-lepton final state is reported. To allow for comparisons with theory calculations, a Bayesian unfolding procedure is applied to account for distortions due to the acceptance and detector effects, leading to partonlevel A C measurements. The data sample at a centre-of-mass energy of 8 TeV, corresponding to an integrated luminosity of 20.3 fb −1 [32], is used to measure A C inclusively and differentially as a function of m tt , p T,tt and β z,tt . This paper is organised as follows. The ATLAS detector is introduced in Sect. 2, followed by the object reconstruction in Sect. 3 and the event selection in Sect. 4. The signal and background modelling is described in Sect. 5 and the procedure to measure A C in Sect. 6. Finally, the results are presented and interpreted in Sect. 7, followed by the conclusions in Sect. 8.

ATLAS detector
The ATLAS detector [33] consists of the following main subsystems: an inner tracking system immersed in a 2 T magnetic field provided by a superconducting solenoid, electromagnetic (EM) and hadronic calorimeters, and a muon spectrometer incorporating three large superconducting toroid magnets composed of eight coils each. The inner detector (ID) is composed of three subsystems: the pixel detector, the semiconductor tracker and the transition radiation tracker. The ID provides tracking information in the pseudorapidity 2 range |η| < 2.5, calorimeters measure energy deposits (clusters) for |η| < 4.9, and the muon spectrometer records tracks within |η| < 2.7. A three-level trigger system [34] is used to select interesting events. It consists of a level-1 hardware trigger, reducing the event rate to at most 75 kHz, followed by two software-based trigger levels, collectively referred to as the high-level trigger, yielding a recorded event rate of approximately 400 Hz on average, depending on the datataking conditions.

Object reconstruction
This measurement makes use of reconstructed electrons, muons, jets, b-jets and missing transverse momentum. A brief summary of the main reconstruction and identification criteria applied for each of these objects is given below. 2 The pseudorapidity is defined in terms of the polar angle θ as η = − ln tan(θ/2) and transverse momentum and energy are defined relative to the beam line as p T = p sin θ and E T = E sin θ. The angular distances are given in terms of R = ( η) 2 + ( φ) 2 , where φ is the azimuthal angle around the beam pipe.
Electron candidates are reconstructed from clusters in the EM calorimeter that are matched to reconstructed tracks in the inner detector. They are required to have a transverse energy, E T , greater than 25 GeV and |η cluster | < 2.47, where η cluster is the pseudorapidity of the electromagnetic energy cluster in the calorimeter with respect to the geometric centre of the detector. Candidates are required to satisfy the tight quality requirements [35] and are excluded if reconstructed in the transition region between the barrel and endcap sections of the EM calorimeter, 1.37 < |η cluster | < 1.52. They are also required to originate less than 2 mm along the z-axis (longitudinal impact parameter) from the selected event primary vertex (PV) 3 and to satisfy two isolation criteria. The first one is calorimeter-based and consists of a requirement on the transverse energy sum of cells within a cone of size R = 0.2 around the electron direction. The second one is a track-based isolation requirement made on the track transverse momentum ( p T ) sum around the electron in a cone of size R = 0.3. In both cases, the contribution from the electron itself is excluded and the isolation cuts are optimised to individually result in a 90 % efficiency for prompt electrons from Z → e + e − decays.
Muon candidates [36,37] are reconstructed using the combined information from the muon spectrometer and the inner detector. They are required to satisfy p T > 25 GeV and |η| < 2.5 and analogously to electrons, the muon track longitudinal impact parameter with respect to the PV is required to be less than 2 mm. Muons are required to satisfy a p T -dependent track-based isolation: the scalar sum of the track p T within a cone of variable size around the muon, R = 10 GeV/p μ T (excluding the muon track itself) must be less than 5 % of the muon p T ( p μ T ), corresponding to a 97 % selection efficiency for prompt muons from Z → μ + μ − decays.
Jets are reconstructed with the anti-k t algorithm [38-40] with a radius parameter R = 0.4 from calibrated topological clusters [33] built from energy deposits in the calorimeters. Prior to jet finding, a local cluster calibration scheme [41,42] is applied to correct the topological cluster energies for the effects of the noncompensating response of the calorimeter, dead material and out-of-cluster leakage. The corrections are obtained from simulations of charged and neutral particles and validated with data. After energy calibration [43], jets are required to have p T > 25 GeV and |η| < 2.5. Jets from additional simultaneous pp interactions (pileup) are suppressed by requiring that the absolute value of the jet vertex fraction (JVF) 4 for candidates with p T < 50 GeV and |η| < 2.4 is above 0.5 [44]. All highp T electrons are also reconstructed 3 The method of selecting the PV is described in Sect. 4. 4 The jet vertex fraction is defined as the fraction of the total transverse momentum of the jet's associated tracks that is contributed by tracks from the PV. as jets, so the closest jet within R = 0.2 of a selected electron is discarded to avoid double counting of electrons as jets. Finally, if selected electrons or muons lie within R = 0.4 of selected jets, they are discarded.
Jets are identified as originating from the hadronisation of a b-quark (b-tagged) via an algorithm that uses multivariate techniques to combine information from the impact parameters of displaced tracks as well as topological properties of secondary and tertiary decay vertices reconstructed within the jet [45,46]. The algorithm's operating point used for this measurement corresponds to 70 % efficiency to tag b-quark jets, a rejection factor for light-quark and gluon jets of ∼130 and a rejection factor of ∼5 for c-quark jets, as determined for jets with p T > 20 GeV and |η| < 2.5 in simulated tt events.
The missing transverse momentum (with magnitude E miss T ) is constructed from the negative vector sum of all calorimeter energy deposits [47]. The ones contained in topological clusters are calibrated at the energy scale of the associated highp T object (e.g. jet or electron). The topological cluster energies are corrected using the local cluster calibration scheme discussed in the jet reconstruction paragraph above. The remaining contributions to the E miss T are called unclustered energy. In addition, the E miss T calculation includes contributions from the selected muons, and muon energy deposits in the calorimeter are removed to avoid double counting.

Event selection
Only events recorded with an isolated or non-isolated singleelectron or single-muon trigger under stable beam conditions with all detector subsystems operational are considered.
The triggers have thresholds on p T , the transverse momentum (energy) of the muon (electron). These thresholds are 24 GeV for isolated single-lepton triggers and 60 (36) GeV for non-isolated single-electron (single-muon) triggers. Events satisfying the trigger selection are required to have at least one reconstructed vertex with at least five associated tracks of p T > 400 MeV, consistent with originating from the beam collision region in the x-y plane. If more than one vertex is found, the hard-scatter PV is taken to be the one which has the largest sum of the squared transverse momenta of its associated tracks.
Events are required to have exactly one candidate electron or muon and at least four jets satisfying the quality and kinematic criteria discussed in Sect. 3. The selected lepton is required to match, with R < 0.15, the lepton reconstructed by the high-level trigger. Events with additional electrons satisfying a looser identification criteria based on a likelihood variable [48] are rejected in order to suppress di-leptonic backgrounds (tt or Z +jets). At this point, the events are sep-arated into three signal regions defined by the number of b-tagged jets (zero, one and at least two).
In order to further suppress multijet and Z +jets backgrounds in events with exactly zero or one b-tagged jets, the following requirements on E miss T and m W T 5 are applied: m W T + E miss T > 60 GeV for events with exactly zero or one btagged jets, and E miss T > 40 (20) GeV for events with exactly zero (one) b-tagged jets.
After the event selection, the main background is the production of W +jets events. Small contributions arise from multijet, single top quark, Z +jets and diboson (W W, W Z, Z Z) production. For events with exactly one (at least two) b-tagged jet(s), 216,465 (193,418) data events are observed, of which 68 % (89 %) are expected to be tt.

Signal and background modelling
Monte Carlo simulated samples are used to model the tt signal and all backgrounds except for those from multijet events, which are estimated from data. All simulated samples utilise Photos (version 2.15) [49] to simulate photon radiation and Tauola (version 1.20) [50] to simulate τ decays. They also include simultaneous pp interactions (pile-up), generated using Pythia 8. 1 [51], and reweighted to the number of interactions per bunch crossing in data (on average 21 in 2012). Most of them are processed through a full Geant4 [52] simulation of the detector response [53], and only the alternative tt samples described in Sect. 5.1 are produced using the ATLAS fast simulation that employs parameterised showers in the calorimeters [54]. Finally, the simulated events are reconstructed using the same software as the data. Further details on the modelling of the signal and each of the backgrounds are provided below.

tt signal
The default simulated tt events are generated with the NLO generator Powheg-Box (version 1, r2330) [55-57] using the CT10 PDF set [58] interfaced to Pythia (version 6.427) [59] with the CTEQ6L1 PDF set and the Peru-gia2011C set of tunable parameters (tune) [60] for the underlying event (UE). The h damp factor, which is the model parameter that controls matrix element/parton shower matching in Powheg-Box and effectively regulates the highp T radiation, is set to the top-quark mass.
The alternative samples used to study the modelling of tt are: where p T is the transverse momentum (energy) of the muon (electron) and φ is the azimuthal angle separation between the lepton and the direction of the missing transverse momentum.
• Mc@nlo (version 4.01) [61] using the CT10 PDF set and interfaced to Herwig (version 6.520) [62] and Jimmy (version 4.31) [63]. • Powheg-Box using the CT10 PDF and setting the h damp parameter to infinity, interfaced to Pythia (version 6.426) with the CTEQ6L1 PDF set and the Perugia2011C UE tune. • Powheg-Box using the CT10 PDF and setting the h damp parameter to infinity, and interfaced to Herwig with the CTEQ6L1 PDF set and Jimmy to simulate the UE. • AcerMC [64] using the CTEQ6L1 PDF set and interfaced to Pythia (version 6.426).

W/Z +jets background
Samples of events with a W or Z boson produced in association with jets (W/Z +jets) are generated with up to five additional partons using the Alpgen (version 2.14) [72] LO generator and the CTEQ6L1 PDF set, interfaced to Pythia (version 6.426) for parton showering and fragmentation. To avoid double counting of partonic configurations generated by both the matrix-element calculation and the parton shower, a parton-jet matching scheme ("MLM matching") [73] is employed. The W +jets samples are generated separately for W +light-jets, W bb+jets, W cc+jets, and W c+jets. The Z +jets samples are generated separately for Z +light-jets, Zbb+jets, and Zcc+jets. Overlap between W/Z QQ+jets (Q = b, c) events generated from the matrix-element calculation and those generated from parton-shower evolution in the W/Z +light-jets samples is avoided via an algorithm based on the angular separation between the extra heavy quarks: if R(Q,Q) > 0.4, the matrix-element prediction is used, otherwise the parton-shower prediction is used. The Z +jets background is normalised to its inclusive NNLO theoretical cross section [74], while data is used to normalise W +jets (see below for details). Further corrections are applied to Z +jets simulated events in order to better describe data in the preselected sample. A correction to the heavyflavour fraction was derived to reproduce the relative rates of Z +2-jets events with zero and one b-tagged jets observed in data. In addition, the Z boson p T spectrum was compared between data and the simulation in Z +2-jets events, and a reweighting function was derived in order to improve the modelling as described in Ref. [75].
The procedure to estimate the normalisation of the W +jets background in data exploits the difference in production cross section at the LHC between W + and W − , where the W + production cross section is higher than W − [76]. This is due to the higher density of u quarks in protons with respect to d quarks, which causes more ud → W + to be produced than dū → W − . The W boson charge asymmetry is then defined as the difference between the numbers of events with a single positive or negative lepton divided by the sum. The prediction for the W boson charge asymmetry in W +jets production is little affected by theoretical uncertainties and can be exploited, in combination with constraints from W + and W − data samples, to derive the correct overall normalisation for the MC sample prediction. The W boson charge asymmetry depends on the flavour composition of the sample, as the size and sign of the asymmetry varies for W bb+jets, W cc+jets, W c+jets, and W +light-jets production. The in situ calibration procedure embedded in the unfolding and described in Sect. 6.4, uses different signal and control regions to determine the normalisation of the W +jets background.

Multijet background
Multijet events can enter the selected data sample through several production and misreconstruction mechanisms. In the electron channel, the multijet background consists of non-prompt electrons from heavy-flavour decays or photon conversion or jets with a high fraction of their energy deposited in the EM calorimeter. In the muon channel, the background contributed by multijet events is predominantly due to final states with non-prompt muons, such as those from semileptonic b-or c-hadron decays. The multijet background normalisation and shape are estimated from data using the "Matrix Method" (MM) technique.
The MM exploits differences in the properties used for lepton identification between prompt, isolated leptons from W and Z boson decays (referred to as "real leptons") and those where the leptons are either non-isolated or result from the misidentification of photons or jets (referred to as "fake leptons"). For this purpose, two samples are defined after imposing the event selection described in Sect. 4, differing only in the lepton identification criteria: a "tight" sample and a "loose" sample, the former being a subset of the latter. The tight selection employs the final lepton identification criteria used in the analysis. For the loose selection, the lepton isolation requirements are omitted for both the muon and electron channels, and the quality requirements are also loosened for the electron channel. The method assumes that the number of selected events in each sample (N loose and N tight ) can be expressed as a linear combination of the numbers of events with real and fake leptons, so that the number of multijet events in the tight sample is given by where real ( fake ) represents the probability for a real (fake) lepton that satisfies the loose criteria to also satisfy the tight. Both of these probabilities are measured in data control samples. To measure real , samples enriched in real leptons from W boson decays are selected by requiring high E miss T or transverse mass m W T . The average real is 0.75 (0.98) in the electron (muon) channel. To measure fake , samples enriched in multijet background are selected by requiring either low E miss T (electron channel) or high transverse impact parameter significance for the lepton track (muon channel). The average fake value is 0.35 (0.20) in the electron (muon) channel. Dependencies of real and fake on quantities such as lepton p T and η, R between the lepton and the closest jet, or number of b-tagged jets, are parameterised in order to obtain a more accurate estimate.

Other backgrounds
Samples of single-top-quark backgrounds corresponding to the t-channel, s-channel and W t production mechanisms are generated with Powheg-Box (version 3.0) [77,78] using the CT10 PDF set. All samples are generated assuming a top-quark mass of 172.5 GeV and are interfaced to Pythia (version 6.425) with the CTEQ6L1 PDF set and the Peru-gia2011C UE tune. Overlaps between the tt and W t final states are removed using the "diagram removal" scheme [79]. The single-top-quark samples are normalised to the approximate NNLO theoretical cross sections [80][81][82] using the MSTW 2008 NNLO PDF set.
Most of the diboson W W/W Z/Z Z+jets samples are generated using Alpgen (version 2.13), with up to three additional partons, and using the CTEQ6L1 PDF set, interfaced to Herwig and Jimmy (version 4.31) for parton showering, fragmentation and UE modelling. For the W W +jets samples, it is required that at least one of the W bosons decays leptonically, while for the W Z/Z Z+jets samples, it is demanded that at least one of the Z bosons decays leptonically. Additional samples of W Z+jets, requiring the W and Z bosons to decay leptonically and hadronically, respectively, are generated with up to three additional partons, including massive b-and c-quarks, using Sherpa v1.4.1 [83] and the CT10 PDF set. All diboson samples are normalised to their NLO theoretical cross sections [84].

Charge asymmetry measurement
To measure the charge asymmetry in top-quark pair events, the full tt system is reconstructed (Sect. 6.1) and the |y| spectra are unfolded to measure parton-level charge asymmetries (Sect. 6.2) using the estimation of the backgrounds and systematic uncertainties (Sect. 6.3). Significant improvements to the analysis method with respect to the 7 TeV measurement [4] have been made, and these improvements are detailed in the description of the measurement in Sect. 6.4.

Reconstruction of the tt kinematics
The reconstruction of the tt system is achieved using a kinematic fit [85] that assesses the compatibility of the observed event with the decays of a tt pair based on a likelihood approach. The basic reconstruction method is explained in Ref. [86], but some modifications are introduced as discussed in the following paragraph.
In events with four or five jets, all jets are considered in the fit. For events where more than five jets are reconstructed, only the two jets with the highest likelihood to be b-jets, according to the multivariate selection (see Sect. 3), and, of the remaining jets, the three with the highest p T are considered in the fit. This selection of input jets for the likelihood was chosen to optimise the correct-sign fraction of reconstructed |y|. The average correct-sign fraction is estimated with simulation studies and found to be 72 and 75 % in events with exactly one and at least two b-tagged jets, respectively. The most probable combination out of all the possible jet permutations is chosen. Permutations with non-b-tagged jets assigned as b-jets and vice versa have a reduced weight due to the tagging probability in the likelihood. Finally, the lepton charge Q is used to determine if the reconstructed semileptonically-decaying quark is a top quark (Q > 0) or an anti-top quark (Q < 0). The distributions of reconstructed quantities, m tt , p T,tt and β z,tt are shown in Fig. 1, with the binnings that are used in the differential measurements.

Unfolding
The reconstructed |y| distributions are distorted by acceptance and detector resolution effects. An unfolding procedure is used to estimate the true |y| spectrum, as defined by the t andt after radiation and before decay in Monte Carlo events, from the one measured in data. The observed spectrum is unfolded using the fully Bayesian unfolding (FBU) technique [87].
The FBU method consists of the strict application of Bayesian inference to the problem of unfolding. This application can be stated in the following terms: given an observed spectrum D with N r reconstructed bins, and a response matrix M with N r × N t bins giving the detector response to a true spectrum with N t bins, the posterior probability density of the true spectrum T (with N t bins) follows the probability density where L ( D|T ) is the likelihood function of D given T and M, and π (T ) is the prior probability density for T . While the response matrix is estimated from the simulated sample of tt events, a uniform prior probability density in all bins is chosen as π (T ), such that equal probabilities to all T spectra within a wide range are assigned. The unfolded asymmetry The treatment of systematic uncertainties is consistently included in the Bayesian inference approach by extending the likelihood L ( D|T ) with nuisance parameter terms. The marginal likelihood is defined as where θ are the nuisance parameters, and N (θ ) their prior probability densities, which are assumed to be Normal distributions with mean μ = 0 and standard deviation σ = 1. A nuisance parameter is associated with each of the uncertainty sources (as explained below).
The marginalisation approach provides a natural framework to treat simultaneously the unfolding and background estimation using multiple data regions. Given the distributions D i measured in N ch independent channels, the likelihood is extended to the product of likelihoods of each channel, so that (6) where the nuisance parameters are common to all analysis channels.

Systematic uncertainties
Several sources of systematic uncertainty are considered, which can affect the normalisation of signal and background and/or the shape of the relevant distributions. Individual sources of systematic uncertainty are considered to be uncorrelated. Correlations of a given systematic uncertainty with others are maintained across signal and background processes and channels. The following sections describe each of the systematic uncertainties considered in the analysis. Experimental uncertainties and background modelling uncertainties (Sects. 6.3.1, 6.3.2) are marginalised during the unfolding procedure, while signal modelling uncertainties, uncertainties due to Monte Carlo sample size, PDF uncertainties and unfolding response uncertainties (Sects. 6.3.3, 6.3.4) are added in quadrature to the unfolded uncertainty. sis. The jet energy resolution (JER) has been determined as a function of jet p T and rapidity using dijet events from data and simulation. The JER in data and in simulation are found to agree within 10 %, and the corresponding uncertainty is assessed by smearing the jet p T in the simulation. The JES and JER uncertainties represent the leading sources of uncertainty associated with reconstructed objects in this analysis.
Heavy-and light-flavour tagging: The efficiencies to tag jets from b-quarks, c-quarks, and light quarks are measured in data as a function of p T (and η for light-quark jets), and these efficiencies are used to adjust the simulation to match data. The uncertainties in the calibration are propagated through this analysis and represent a minor source of uncertainty.

Jet reconstruction and identification:
The uncertainty associated with the jet reconstruction efficiency is assessed by randomly removing 0.2 % of the jets with p T below 30 GeV, to match the measured jet inefficiency in data for this p T range [43]. The uncertainty on the efficiency that each jet satisfies the JVF requirement is estimated by changing the JVF cut value from its nominal value by ±0.1, and repeating the analysis using the modified cut value. Both uncertainties have a negligible impact on the measurement.
The combined effect of all these uncertainties results in an overall normalisation uncertainty on the signal and background of approximately 1.5 %. Charge misidentification is not considered as it is small [88] and has a negligible impact on the measurement.

Missing transverse momentum:
The E miss T reconstruction is affected by uncertainties associated with leptons, jet energy scales and resolutions which are propagated to the E miss T calculation. Additional small uncertainties associated with the modelling of the underlying event, in particular its impact on the p T scale and resolution of unclustered energy, are also taken into account. All uncertainties associated with the E miss T have a negligible effect.

Luminosity:
The uncertainty on the integrated luminosity is 2.8 %, affecting the overall normalisation of all processes estimated from MC simulation. It is derived following the methodology detailed in Ref. [32]. The impact of this uncertainty is negligible in this measurement.

Background modelling
W+jets: The predictions of normalisation and flavour composition of the W +jets background are affected by large uncertainties, but the in situ data-driven technique described in Sect. 5.2 reduces these to a negligible level. All sources of uncertainty other than normalisation are propagated to the W +jets estimation.
Z+jets: Uncertainties affecting the modelling of the Z +jets background include a 5 % normalisation uncertainty from the theoretical NNLO cross section [74], as well as an additional 24 % normalisation uncertainty added in quadrature for each additional inclusive jet-multiplicity bin, based on a comparison among different algorithms for merging LO matrix elements and parton showers [89]. The normalisation uncertainties for Z +jets are described by three uncorrelated nuisance parameters corresponding to the three b-tag multiplicities considered in the analysis.
Multijet background: Uncertainties on the multijet background estimated via the Matrix Method receive contributions from the size of the data sample as well as from the uncertainty on fake , estimated in different control regions. A normalisation uncertainty of 50 % due to all these effects is assigned independently to the electron and muon channels and to each b-tag multiplicity, leading to a total of six uncorrelated uncertainties.
Other physics backgrounds: Uncertainties affecting the normalisation of the single-top-quark background include a +5 %/-4 % uncertainty on the total cross section estimated as a weighted average of the theoretical uncertainties on t-, W t-and s-channel production [80][81][82]. Including an additional uncertainty in quadrature of 24 % per additional jet has a negligible impact on the measurement. Uncertainties on the diboson background normalisation include 5 % from the NLO theoretical cross sections [84] added in quadrature to an uncertainty of 24 % due to the extrapolation to the high jet-multiplicity region, following the procedure described for Z +jets.

Signal modelling
In order to investigate the impact of uncertainties on the tt signal modelling, additional samples generated with Powheg-Box interfaced to Herwig, Mc@nlo interfaced to Herwig and AcerMC interfaced to Pythia are considered (see Sect. 5.1 for more details). Different predictions and response matrices built with those tt samples are used to repeat the full analysis procedure isolating one effect at the time. For each case, the intrinsic asymmetry and the unfolded asymmetry are measured. The intrinsic asymmetry is the asymmetry generated in each Monte Carlo sample before the simulation of the detector response. Double differencees between the intrinsic (int) asymmetry and the unfolded (unf) values of the nominal (nom) and the alternative (alt) sample are considered as uncertainties to account for the different A C predictions of the different samples, . This is referred to as the double difference.

NLO generator:
The uncertainty associated with the choice of NLO generator is estimated from the double difference of the parton-level A C and unfolded A C comparing Powheg-Box interfaced to Herwig (nom) and Mc@nlo interfaced to Herwig (alt).

Fragmentation model:
The uncertainty associated with the fragmentation model is estimated from the double difference of the parton-level A C and unfolded A C comparing Powheg-Box interfaced to Pythia (nom) and Powheg-Box interfaced to Herwig (alt).

Initial-and final-state radiation (ISR/FSR):
The uncertainty associated with the ISR/FSR modelling is estimated using the AcerMC generator where the parameters of the generation were varied to be compatible with the results of a measurement of tt production with a veto on additional central jet activity [90]. Two variations producing more and less ISR/FSR are considered. The uncertainty is estimated from half of the double difference of the parton-level A C and unfolded A C comparing Powheg-Box (nom) and AcerMC (alt) interfaced to Pythia producing more and less ISR/FSR.

Others
Monte Carlo sample size: To assess the effect on the measurement of the limited number of Monte Carlo events, an ensemble of 1000 response matrices, each of them fluctuated according to the raw number of simulated events, is produced. Unfolding is repeated with the same pseudo-dataset for each fluctuated response matrix. The uncertainty is estimated as the standard deviation of the ensemble of the 1000 A C values obtained. The estimated systematic uncertainty associated with limited number of Monte Carlo events is about ten times smaller than the data statistical uncertainty; this is consistent with the size of the available Monte Carlo sample.

PDF uncertainties:
The choice of PDF in simulation has a significant impact on the charge asymmetry of the simulated W +jets background. Since this asymmetry is exploited to calibrate the W +jets prediction, the related uncertainty has to be estimated. The uncertainty on the PDFs is evaluated using three different PDF sets: CT10 [58], MSTW 2008 [91] and NNPDF2.1 [92]. For each set, the PDFs are varied based on the uncertainties along each of the PDF eigenvectors. Each variation is applied by reweighting the W +jets sample eventby-event. The A C measurements are repeated for each varied W +jets template and the uncertainty is estimated as half of the largest difference between any variation of CT10 and MSTW 2008, and the ±1σ variations for NNPDF2.1. The resulting uncertainties are small, but non-negligible. The impact of uncertainties related to PDFs are found to be negligible in tt modelling.
Unfolding response: The response of the unfolding procedure, i.e. any non-linearity or bias, is determined using a set of six pseudo-datasets, each of them being composed of the default tt signal reweighted to simulate an asymmetry and the default MC simulation predictions. The injected A C value ranges between −0.2 and 0.2 depending on the differential variable and bin. The six reweighted pseudo-datasets are unfolded using the default response matrix and the uncertainty associated with the unfolding response is calculated as: , with a and b the slope and offset of a linear fit of the generator-level (intrinsic) A C versus unfolded A C of the six reweighted pseudo-datasets previously defined and A meas C the measured value in data.

Measurement
A fit is performed which maximises the extended likelihood of Eq. (6). In this fit, the events are further separated based on the sign of the lepton charge Q . The measurements are then performed using a combination of six channels based on the lepton charge (Q > 0 and Q < 0) and the b-jet multiplicity (zero b-jets, one b-jet, at least two b-jets). The |y| distribution is split into four bins in all the channels except the zero b-jets channel, as no extra information for A C is expected. Four bins in |y| are considered in each differential bin of all differential measurements.
The W +jets in situ calibration procedure consists of fitting the calibration factors K bb/cc , K c and K light for scaling the flavor components of the W +jets background with different charge asymmetries, assuming uniform prior probabilities π during the posterior probability estimation defined in Eq. (7). The b-jet multiplicity provides information about the heavyand light-flavour composition of the W +jets background, while the lepton charge asymmetry is used to determine the normalisation of each component. Figure 2 shows the different W +jets contributions for the different b-jet multiplicities and lepton charges. In addition to the expected number of tt events for each bin in T , the W +jets calibration factors are free parameters in the likelihood. The posterior probability density is thus where B = B(K bb/cc , K c , K light ; θ s , θ b ) is the total background prediction, the probability densities π are uniform priors and R is the reconstructed signal prediction. Two cat-

Fig. 2
Comparison between prediction and data for the 18 bins used in the inclusive A C measurement before (top) and after (bottom) the simultaneous unfolding procedure and W +jets in situ background calibration, including only uncertainties that are marginalized. The |y| distribution in four bins is considered for the tt-enriched event samples with exactly one and at least two b-jets; a single bin is considered for the background-enriched sample with zero b-jets. After the calibration, the background components are scaled to the measured values for the nuisance parameters, and the prediction for tt events in each bin is estimated by folding the measured parton-level parameters through the response matrix. The bottom part of each plot shows the ratio of the data to the predicted value together with combined statistical and systematic uncertainties egories of nuisance parameters are considered: the normalisation of the background processes (θ b ), and the uncertainties associated with the object identification, reconstruction and calibration (θ s ). While the first ones only affect the background predictions, the latter, referred to as object systematic uncertainties, affect both the reconstructed distribution for tt signal and the total background prediction. The W +jets calibration factors are found to be K bb/cc = 1.50 ± 0.11, K c = 1.07±0.27 and K light = 0.80±0.04, where the uncertainties include both the statistical and systematic components.
The final numbers of expected and observed data events after the full event selection, marginalisation of nuisance parameters and W +jets in situ calibration are listed in Table 1, while Fig. 2 shows the good level of agreement between the data and expectation before and after marginalisation for the six channels. In both cases, the uncertainties that are marginalized are shown. Since these uncertainties are correlated for the background and signal components, the total combined marginalized uncertainty is smaller than the sum of the constituent parts.
Since the background estimation is part of the Bayesian inference procedure described in Sect. 6.2, it is not possible to study the impact of systematic uncertainties by repeating unfolding on data with varied templates, without using marginalisation. Instead, the expected impact of systematic uncertainties is studied with pseudo-data distributions corresponding to the sum of the background and signal predictions. For each source of uncertainty, the ±1σ variations of the predictions are used to build the pseudodata, and the unfolding procedure is repeated. The baseline background templates and response matrices, as in the actual measurements, are used. Table 2 shows the average asymmetry variation δ A C computed, for each source of uncertainty, as |A C (+1σ ) − A C (−1σ )|/2, but only the uncertainties having a variation above 10 % of the statistical uncertainty are reported in the table. The total uncertainty associated with the marginalised systematic uncertainties is estimated by subtracting in quadrature the statis- Table 2 Impact of individual sources of uncertainty on the inclusive A C measurement. All uncertainties described in Sect. 6.3 are considered, but only the ones having a variation above 10% of the statistical uncertainty are reported in the table. tical term from the total marginalised uncertainty. It yields 0.002 (category (a) in Table 2). The total, non-marginalised uncertainty associated with systematic uncertainties is estimated by summing in quadrature sources from category (b) in Table 2.
The precision of the measurement is limited by the statistical uncertainty, and the main sources of systematic uncertainty are the signal modelling and the uncertainties with a large impact on the size of the W +jets background, such as the uncertainty on the jet energy scale and resolution.

Differential measurements
The A C differential spectra are compared in Fig. 3 with the theoretical SM predictions, as well as with BSM predictions for right-handed colour octets with low and high masses [93]. The BSM predictions are not shown in the measurement as a function of p T,tt as they are LO 2 → 2 calculations. The results are compatible with the SM, and it is not possible to distinguish between the SM and BSM models at this level of Fig. 3 Measured A C values as a function of bin-averaged m tt , β z,tt and p T,tt , compared with predictions for SM [1] and for right-handed colour octets with masses below the tt threshold and beyond the kine-matic reach of current LHC searches [93]. The BSM predictions are shown only for the two top plots. The bins are the same as the ones reported in Tables 3 and 4 precision. The BSM models are tuned to be compatible with the Tevatron asymmetry measurements and the A C measurements at √ s = 7 TeV. Table 3 shows the average asymmetry variation δ A C computed for each differential measurement, for each source of uncertainty, as explained in Sect. 7.1. The precision of the Table 3 Impact of individual sources of uncertainty on the measurement of A C in bins of m tt , β z,tt and p T,tt . All uncertainties described in Sect. 6.3 are considered, but only the ones having at least one bin with a variation above 10 % of the statistical uncertainty are reported in the table; the others are quoted as "-".    Table 4 for the differential measurements as a function of m tt β z,tt and p T,tt . The theoretical values are described in Ref. [1] (SM) and Ref. [93] (BSM), and they have been provided for the chosen bins. The correlation matrices are shown in Table 5 for the measurements as a function of m tt , β z,tt and p T,tt .
In regions with sensitivity to BSM (high values of m tt and β z,tt ), the uncertainty on the measurements is largely dominated by the available statistics, while in other regions the uncertainty on signal modeling and/or parton shower dominates. Figure 4 shows the inclusive A C measurement presented in Sect. 7. The measurement is compared to the tt forward-backward asymmetry 6 A FB measured at the Tevatron by CDF and D0 experiments. Predictions given by several BSM models, the details of which can be found in Refs. [20,94], are also displayed. These BSM models include a W boson, a heavy axigluon (G μ ), a scalar isodoublet (φ), a colourtriplet scalar (ω 4 ), and a colour-sextet scalar ( 4 ). For each model, the predictions for A FB and A C are derived using the PROTOS generator [95] with the constraints described in Ref. [86]. The ranges of predicted values for A FB and A C for a given set of BSM model are also shown. The BSM physics contributions are computed using the treelevel SM amplitude plus the one(s) from the new particle(s), to account for the interference between the two contributions. The phase-space of the parameters describing the various BSM models (such as the BSM particle masses and couplings) is limited by the measurement presented in this paper.  Fig. 4 Measured inclusive charge asymmetries A C at the LHC versus forward-backward asymmetries A FB at Tevatron, compared with the SM predictions [1,9] as well as predictions incorporating various potential BSM contributions [20,94]: a W boson, a heavy axigluon (G μ ), a scalar isodoublet (φ), a colour-triplet scalar (ω 4 ), and a colour-sextet scalar ( 4 ). The horizontal bands and lines correspond to the ATLAS and CMS measurements, while the vertical ones correspond to the CDF and D0 measurements. The uncertainty bands correspond to a 68 % confidence level interval. The figure on the right is a zoomed-in version of the figure on the left

Conclusion
The top-quark pair production charge asymmetry was measured with pp collisions at the LHC using an integrated luminosity of 20.3 fb −1 recorded by the ATLAS experiment at a centre-of-mass energy of √ s = 8 TeV in tt events with a single lepton (electron or muon), at least four jets and large missing transverse momentum. The reconstruction of tt events was performed using a kinematic fit. The reconstructed inclusive distribution of |y| and the distributions as a function of m tt , p T,tt and β z,tt were unfolded to obtain results that can be directly compared to theoretical computations. The measured inclusive tt production charge asymmetry is A C = 0.009 ± 0.005 (stat.+ syst.), to be compared to the SM prediction A C = 0.0111 ± 0.0004 [1]. All measurements presented in this paper are statistically limited and are found to be compatible with the SM prediction within the uncertainties. The precision of the measurements also allows for the exclusion of a large phase-space of the parameters describing various BSM models.