UvA-DARE (Digital Academic Measurement of single top-quark production in association with a W boson in the single-lepton channel at √s = 8 TeV with the ATLAS detector

The production cross-section of a top quark in association with a W boson is measured using proton–proton collisions at √ s = 8 TeV. The dataset corresponds to an integrated luminosity of 20 . 2 fb − 1 , and was collected in 2012 by the ATLAS detector at the Large Hadron Collider at CERN. The analysis is performed in the single-lepton chan-nel.Eventsareselectedbyrequiringoneisolatedlepton(elec- tron or muon) and at least three jets. A neural network is trained to separate the tW signal from the dominant t ¯ t background. The cross-section is extracted from a binned proﬁle maximum-likelihood ﬁt to a two-dimensional discriminant built from the neural-network output and the invariant mass of the hadronically decaying W boson. The measured cross-section is σ tW = 26 ± 7 pb, in good agreement with the Standard Model expectation.


Introduction
Single top quarks are produced in proton-proton collisions via the weak interaction. At leading order (LO) three different channels, which depend on the virtuality of the W boson involved, are defined: t-channel, s-channel or top-quark production in association with a W boson, called tW production. These processes, for which example Feynman diagrams are shown in Fig. 1, involve a W tb vertex at LO in the Standard Model (SM). Calculations involving tW production beyond LO have to include quantum interference with tt production. Measurements of single-top-quark cross-sections are used to study the properties of this vertex, as they are directly sensitive to the Cabibbo-Kobayashi-Maskawa (CKM) matrix element |V tb |. Deviations from the cross-sections predicted by the SM can originate from single top quarks produced with similar kinematics in the decays of unknown heavy particles predicted by physics beyond the Standard Model. If the masses of these particles are beyond the reach of direct searches, they might be revealed through their effects on the effective W tb coupling [1]. Using measurements in all three channels of single top-quark production, physics beyond the SM can be probed systematically in the context of Effective Field Theory [2]. As each of the single-top-quark processes can be sensitive to different sources of new physics, it is also important to study each channel separately. In addition, the SM production of tW is an important background in direct searches for particles beyond the SM [3,4].
At the Large Hadron Collider (LHC), evidence for the tW production process was found by the ATLAS [5] and CMS Collaborations [6] at √ s = 7 TeV and the process was observed by both experiments [7,8] at √ s = 8 TeV. The tW cross-section has been also measured with 13 TeV collision data inclusively by the CMS Collaboration [9] as well as inclusively and differentially by the ATLAS Collaboration [10][11][12]. These measurements were performed in final states with two leptons, and the measured cross-sections agree with the theoretical expectations. This paper presents evidence for tW production in final states with a single lepton using proton-proton ( pp) collisions at √ s = 8 TeV. This topology features a W boson in addition to a top quark, which decays mainly into another W boson and b-quark, leading to a W + W − b state. In the singlelepton channel, one of the W bosons decays leptonically (W L ) while the other one decays hadronically (W H ). Therefore, the experimental signature of event candidates is characterised by one isolated charged lepton (electron or muon), large missing transverse momentum (E miss T ), and three jets with high transverse momentum ( p T ), one of which contains a b-hadron and is labelled as a b-tagged jet, j B . In contrast to the dilepton analyses, the event signature contains only one neutrino, which originates from the leptonic W -boson decay. Hence, both the W -boson and the top-quark kinematics can be reconstructed and used to separate the signal from background. The main backgrounds are W + jets and tt events, where the latter background poses a major challenge in this measurement because of its similar kinematics and a ten times larger cross-section compared to the tW signal. An artificial neural network is trained to separate the signal from the tt background. The cross-section is extracted using a binned profile maximum-likelihood fit to a two-dimensional discriminant. This measurement, performed with tW single-lepton events, constitutes a crosscheck of the previous results published in the dilepton channel.

ATLAS detector
The ATLAS experiment [13] at the LHC is a multipurpose particle detector with a forward-backward symmetric cylindrical geometry and a near 4π coverage in solid angle. 1 It consists of an inner tracking detector (ID) surrounded 1 ATLAS uses a right-handed coordinate system with its origin at the nominal interaction point (IP) in the centre of the detector and the zaxis along the beam pipe. The x-axis points from the IP to the centre of the LHC ring, and the y-axis points upwards. Cylindrical coordinates by a thin superconducting solenoid providing a 2 T axial magnetic field, electromagnetic and hadron calorimeters, and a muon spectrometer (MS). The ID provides chargedparticle tracking in the pseudorapidity range |η| < 2.5. It consists of silicon pixel, silicon microstrip, and transitionradiation tracking detectors. Lead/liquid-argon (LAr) sampling calorimeters provide electromagnetic (EM) energy measurements with high granularity. An iron/scintillator-tile hadron calorimeter covers the central pseudorapidity range (|η| < 1.7). The endcap (1.5 < |η| < 3.2) and forward (3.1 < |η| < 4.9) regions are instrumented with LAr calorimeters for measurements of both EM and hadronic energy. The MS surrounds the calorimeters and includes a system of precision tracking chambers (|η| < 2.7) and fast detectors for triggering (|η| < 2.4). The magnet system for the MS consists of three large air-core toroidal magnets with eight superconducting coils. The field integral of the toroids ranges between 2.0 and 6.0 T m across most of the detector. Collisions producing interesting events are selected for storage with the trigger system [14]. For the data taken at √ s = 8 TeV, a three-level trigger system was used to select events. The first-level trigger is implemented in hardware and uses a subset of the detector information. It reduced the accepted rate to at most 75 kHz. This was followed by two software-based trigger levels that together reduced the accepted event rate to 400 Hz on average, depending on the data-taking conditions.

Data and simulated event samples
The data considered in this analysis are from pp collisions at √ s = 8 TeV and were taken with stable LHC beams and Footnote 1 continued (r, φ) are used in the transverse plane, φ being the azimuthal angle around the z-axis. The pseudorapidity is defined in terms of the polar angle θ as η = − ln tan(θ/2). Angular distance is measured in units of the ATLAS detector fully operational, corresponding to an integrated luminosity of 20.2 fb −1 . Monte Carlo (MC) samples were produced using the full ATLAS detector simulation [15] implemented in Geant 4 [16]. In addition, alternative MC samples, used to train the neural network and evaluate systematic uncertainties, were produced using AtlFast2 [15], which provides a faster calorimeter simulation making use of parameterised showers to compute the energy deposited by the particles. Pile-up (additional pp interactions in the same or nearby bunch crossing) was modelled by overlaying simulated minimum-bias events generated with Pythia 8 [17]. Weights were assigned to the simulated events, such that the distribution of the number of pp interactions per bunch crossing in the simulation matches the corresponding distribution in the data, which has an average of 21 [18].
The tW signal events were simulated using the nextto-leading order (NLO) Powheg method [19][20][21] implemented in the Powheg Box (v.1.0) generator (revision 2192) [22] with the CT10 parton distribution function (PDF) set [23] in the matrix-element calculation. The mass and width of the top-quark were set to m t = 172.5 GeV and t = 1.32 GeV, respectively. The top quark was assumed to decay exclusively into W b. The parton shower, hadronisation and underlying event were simulated using Pythia 6 (v6.426) [24] with the LO CTEQ6L1 PDF set [25] and a corresponding set of tuned parameters called the Perugia 2011 (P2011C) tune [26]. The factorisation scale, μ f , and renormalisation scale, μ r , were set to m t . Calculations involving tW production beyond LO included quantum interference with tt production. Double-counting of the contributions was avoided by using either the diagram-removal (DR) or the diagram-subtraction (DS) scheme [27,28]. In the DR scheme, diagrams with a second on-shell top-quark propagator are removed from the amplitude, while in the DS scheme, a subtraction term cancels out the tt contribution to the cross-section when the top-quark propagator becomes on shell. Nominal MC samples were generated using the DR scheme. For the evaluation of systematic uncertainties, alternative samples were generated using the DS scheme, or using Powheg Box or MC@NLO (v4.06) [29], each interfaced with Herwig (v6.520) [30]. For the Herwig samples, the AUET2 tune [31] with the CT10 PDF was used and the underlying event was generated with Jimmy (v4.31) [32]. In addition, Pythia 6 (v6.427) samples with variations of μ r and μ f and the radiation tunes were used. The SM tW crosssection prediction at NLO including next-to-next-to-leadinglog (NNLL) soft gluon corrections [33,34] was calculated as σ th.
tW (8 TeV) = 22.4 ± 0.6 (scale) ± 1.4 (PDF) pb assuming a top-quark mass, m t , of 172.5 GeV. The first uncertainty accounts for renormalisation and factorisation scale variations (from m t /2 to 2m t ) and the second term covers the uncertainty in the parton distribution functions, evalu-ated using the MSTW2008 PDF set [35] at next-to-next-toleading order (NNLO).
The tt sample was generated with Powheg Box (v1.1) interfaced with Pythia 6 (v6.427) [36]. In the Powheg Box event generator, the CT10 PDFs were used, while the CTEQ6L1 PDFs were used for Pythia. The h damp parameter, which effectively regulates the highp T gluon radiation, was set to m t . The predicted tt production cross-section, σ tt (8 TeV) = 252.9 +6.4 −8.6 (scale) ± 11.7 (PDF + α s ) pb, was calculated with the Top++2.0 program to NNLO in perturbative QCD, including soft-gluon resummation to NNLL [37]. The first uncertainty comes from the sum in quadrature of the effects of independently varying μ r and μ f . The uncertainty associated with variations in the PDFs and strong coupling constant, α S , was evaluated following the PDF4LHC NLO prescription [38,39], which defines the central value as the midpoint of the uncertainty envelope of three PDF sets: MSTW2008 NNLO [35], CT10 NNLO [40] and NNPDF2.3 5f FFN [41]. The same procedures as for the tW samples were employed to determine the uncertainties due to the NLO matching method and the parton shower and hadronisation. Samples to evaluate the scale uncertainties were produced in a similar way, varying μ r and μ f together with the Perugia tune, but also adding variations in the h damp parameter (for the up-variation, h damp was changed to 2m t , while for the down variation it was kept at m t ).
The other single-top-quark production processes, s-channel and t-channel, were also generated with Powheg Box (v1.1) coupled to Pythia 6 (v6.426), using the same PDF sets as described for the other top-quark processes above. The predicted cross-sections at √ s = 8 TeV were calculated at NLO plus NNLL as 5.6 ± 0.2 pb for the s-channel [42,43], and 87.8 +3.4 −1.9 pb for the t-channel [44,45] process. The multi-leg LO generator Sherpa (v1.4.1) [46-48], together with the CT10 PDF sets, was used to simulate vector-boson production in association with jets. Sherpa was used to generate the hard process as well as the parton shower and the modelling of the underlying event. Doublecounting between the inclusive V + n parton samples (with V = W or Z ) and samples with associated heavy-quark pair production was avoided consistently by using massive c-and b-quarks in the shower. The predicted NNLO W +jets cross-section with W decaying leptonically was calculated as σ ( p p → ± ν X ) = 36.3 ± 1.9 nb [49]. 2 For Z + jets the cross-section was calculated at NNLO in QCD for leptonic Z decays as σ ( pp → + − X ) = 3.72 ± 0.19 nb [49]. The AtlFast2 simulation was used to generate these samples with sufficient statistics. For cross-checks of the W + jets modelling, an alternative sample generated with Alpgen (v2.14) [50] with up to five additional partons, Pythia 6 (v6.426) and the CTEQ6L1 PDFs were used. Dibo-2 The quoted cross-section is the sum of decays into e, μ and τ . 123 son samples (WW/W Z/Z Z + jets) were generated with Herwig (v6.520) at LO QCD using the CTEQ6L1 PDF. The theoretical NLO cross-section for events with one lepton is 29.4 ± 1.5 pb [51].
Multijet events are selected in the analysis when they contain jets or photons misidentified as leptons or contain non-prompt leptons from hadron decays (both referred to as a 'fake' lepton). This background was estimated directly from data using the matrix method [52], which exploits differences in lepton identification and isolation properties between prompt and fake leptons. The data were processed with a second, 'loose' set of lepton selection criteria. The resulting sample was then corrected for efficiency differences between the two sets of cuts, and the contamination from events containing prompt leptons was subtracted. The efficiencies, lepton selection criteria, and uncertainties applied in this analysis are the same as in Ref. [52].

Object definitions
Primary vertex (PV) candidates in the interaction region are reconstructed from at least five tracks that satisfy a transverse momentum ( p T ) of p T > 400 MeV. The candidate with the highest sum of p 2 T over all associated tracks is chosen as the hard-collision PV [53].
Muon candidates are reconstructed by matching segments or tracks in the MS with tracks found in the ID [54]. The candidates must have p T > 25 GeV and be in the pseudorapidity range |η| < 2.5. The longitudinal impact parameter of the track relative to the hard-collision PV, |z vtx |, is required to be smaller than 2 mm. In order to reject non-prompt muons, an isolation criterion is applied. The isolation variable is defined as the scalar sum of the transverse momenta of all tracks with p T > 1 GeV (excluding the muon track) within a cone of size R = 10 GeV/ p T (μ) around the muon's direction. It is required to be less than 5% of the muon p T . The selection efficiency after this requirement is measured to be about 97% in Z → μ + μ − events.
Electron candidates are reconstructed from energy deposits (clusters) in the EM calorimeter, which match a wellreconstructed track in the ID [55]. Requirements on the transverse and longitudinal impact parameter of |d vtx | < 1 mm and |z vtx | < 2 mm, respectively, are applied. Electron candidates must have energy in the transverse plane E T > 25 GeV and |η cluster | < 2.47, where η cluster denotes the pseudorapidity of the cluster. Clusters in the calorimeter barrel-endcap transition region, 1.37 < |η| < 1.52, are excluded. An isolation requirement based on the deposited transverse energy in a cone of size R = 0.2 around the direction of the electron and the p T sum of the tracks in a cone with R = 0.3 around the same direction is applied. This requirement is chosen to give a nearly uniform selection efficiency of 85% in p T and η, as measured in Z → e + e − events. Electron candidates that share the ID track with a reconstructed muon candidate are vetoed.
Jets are reconstructed using the anti-k t algorithm [56,57] with a radius parameter of R = 0.4 using topological clusters [58], calibrated with the Local Cluster Weighting method [59], as input to the jet finding. The jet energy is further corrected by subtracting the contribution from pile-up events and applying an MC-based and a data-based calibration. The jet vertex fraction (JVF) [60] variable is used to identify the primary vertex from which the jet originated. The JVF criterion suppresses pile-up jets with p T < 50 GeV and |η| < 2.4. To avoid possible overlap between jets and electrons, jets that are closer than R = 0.2 to an electron are removed. Afterwards, remaining electron candidates overlapping with jets within a distance of R = 0.4 are rejected. Finally, muons overlapping with jets within R = 0.4 are removed.
The identification of jets originating from the hadronisation of a b-quark (b-tagging) is based on various algorithms exploiting the long lifetime, high mass and high decay multiplicity of b-hadrons as well as the properties of the b-quark fragmentation. The outputs of these algorithms are combined in a neural network classifier to maximise the b-tagging performance [61]. The choice of b-tagging working point represents a trade-off between the efficiency for identifying b-jets and rejection of other jets. The chosen working point for this analysis corresponds to a b-tagging efficiency of 70%. The corresponding c-quark-jet rejection factor is about 5 and the light-quark-jet rejection factor is about 120. These efficiencies and rejection factors were obtained using simulated tt events. The tagging efficiencies in the simulation are corrected to match the efficiencies measured in data [61].
The p miss T of the event, defined as the momentum imbalance in the plane transverse to the beam axis, is primarily due to neutrinos that escape detection. It is calculated as the negative vector sum of the transverse momenta of the reconstructed electrons, muons, jets and the clusters that are not associated with any of the previous objects (the 'soft term') [62]. Its magnitude is denoted E miss T .

Event selection
Events are required to have a hard-collision primary vertex. They also have to pass a single-lepton trigger requirement [14,63] and contain at least one electron or muon candidate with p T > 30 GeV matched to the lepton that fired the trigger. The electron trigger requires an electron candidate, formed by an EM calorimeter cluster matched with a track, either with E T > 60 GeV or with E T > 24 GeV and additional isolation requirements. The muon trigger requires a muon candidate, defined as a reconstructed track in the muon spectrometer, either with p T > 36 GeV or with p T > 24 GeV and isolation requirements. If there is another lepton candidate with a transverse momentum above 25 GeV, the event is rejected. This lepton veto guarantees orthogonality with respect to the dilepton analysis. The contribution from leptonically decaying τ -leptons is included. In the following, the electron or muon candidate is referred to as the lepton. Events identified as containing jets from cosmic rays or beam-induced backgrounds or due to noise hot spots in the calorimeter are removed. Only jets with p T > 30 GeV and |η| < 2.4 are considered in the analysis. Additionally, a requirement of E miss T > 30 GeV is applied, and the transverse mass 3 of the leptonically decaying W boson must satisfy m T (W L ) > 50 GeV.
In order to perform the measurement and validate the result, selected events are divided into different categories based on the jet and b-tagged jet multiplicities. The region with three jets of which one is b-tagged (3j1b) is called the signal region and is used to extract the tW cross-section. The region with four jets, two of them b-tagged (4j2b), contains a very pure sample of tt events and is used as the tt validation region to check the modelling of this background. Table 1 shows the expected and the observed numbers of events in the signal region after the event selection. All backgrounds except fake leptons, which is estimated using data-driven methods, are normalised to their expected cross-sections. The tW events constitute about 5% of the total number of events. The major backgrounds are tt production with about 58%, and W+jets production with about 28% of the total number of events. The W+jets events are subdivided into heavy flavour (HF), where a W boson is produced in association with b-or c-jets, and light flavour (LF). The total numbers of expected events agree within a few percent with the observed numbers of events.

Separation of signal from background
Differences between signal and background event kinematics are exploited to better separate them. The tt background is inherently difficult to distinguish from the signal, motivating the use of an artificial neural network (NN) implemented in the NeuroBayes framework [64,65]. Detailed information about how the NN is used in single-top-quark analyses can be found in Ref. [66]. The NN input variables are selected such that they contribute significantly to the statistical separation power between signal and background, while avoiding variables that would lead to an increase of the expected uncertainty in the signal cross-section. The observable m(W H ) (Fig. 2) provides very good separation of the signal from 3 The transverse mass is calculated using the momentum of the lepton associated with the W boson, p miss T and the azimuthal angle between the  Fig. 2a. The peak results from events where the two untagged jets are correctly matched to the hadronically decaying W boson. This is less likely to happen for tt events than for tW events due to the higher b-jet multiplicity and the limited b-tagging efficiency. On the other hand, the W +jets background does not feature such a peak since the W boson must decay leptonically for the events to pass the selection. Figure 2b shows the pre-fit distribution of m(W H ), and also demonstrates good pre-fit modelling of the data.

Neural network
The NN is trained using simulated events with the two reconstructed untagged jets matched within R < 0. contain a lepton, only tW , tt and diboson events can have a pair of jets matched to the hadronic W -boson decay. Given that the contribution from diboson production is very small, the background sample used for the training consists entirely of tt events. Following the training procedure mentioned before, the following four variables (ordered by significance) are selected as input for the NN: • the transverse momentum of the tW system, p T (W H W L j B ), divided by the sum of the objects' transverse momenta, where the four-momentum of W L is the sum of the fourmomenta of the electron or muon and the neutrino, and the four-momentum of the neutrino is determined using  The distribution of the NN response is subdivided into eight bins, with the edges placed approximately at the 12.5% quantiles of a 50:50 mixture of tW and tt events. Figure 4a shows the shape of the NN response for the tW and tt processes and Fig. 4b presents the comparison between data and Monte Carlo simulation.

Two-dimensional discriminant
For the two-dimensional discriminant, m(W H ) is used on the abscissa and the NN response on the ordinate of the twodimensional discriminant. Outside of the aforementioned m(W H ) range from 65 to 92.5 GeV, the bins corresponding to different values of the NN response are merged, i.e. the NN response is ignored. The two-dimensional distribution is presented in Fig. 5.
The bins are then rearranged on a one-dimensional axis in column-major order. The resulting one-dimensional distribution is presented in Fig. 6, together with a comparison of the shapes. The first three bins and the last ten bins correspond directly to the bins of m(W H ) below 65 GeV and above 92.5 GeV respectively. In between are four blocks of eight bins, corresponding to the NN output in slices of m(W H ).  All systematic uncertainties in the reconstruction of jets and leptons are propagated to the uncertainty in E miss T . In addition, dedicated uncertainties are assigned to the soft term of the E miss T , which accounts for energy deposits in the calorimeter which are not matched to highp T physics objects [62].
The uncertainty in the integrated luminosity for the data set used in this analysis is 1.9%. It is derived following the methodology detailed in Ref. [18]. This systematic uncertainty is applied to all contributions determined from the MC simulation.
Uncertainties stemming from theoretical models are evaluated using alternative MC samples for tW and tt processes.
The renormalisation and factorisation scales are varied in the matrix element and in the parton shower together with the amount of QCD radiation. Both scales are varied simultaneously in the matrix element and in the parton shower. The variation of both μ r and μ f by a factor of 0.5 is combined with the Perugia 2012radHi tune, while the variation of the scale parameters by a factor of 2.0 is combined with the Perugia 2012radLo tune [26]. This (radiation) uncertainty is considered uncorrelated between the tW and tt processes. The NLO matrix element generator uncertainty is estimated by comparing two NLO matching methods: Powheg Box and MC@NLO, both interfaced with Herwig. The parton shower, hadronisation and underlyingevent systematic uncertainties are computed by comparing Powheg Box with either Pythia or Herwig. These are treated as fully correlated between the tW and tt processes. The uncertainty due to the treatment of the interference effects of the tW and tt processes is evaluated by using the tW DS scheme instead of the DR scheme, both generated using Powheg Box with Pythia. The effect of the PDF uncertainties on the acceptance is taken into account for both the tW signal and the tt background and treated as uncorrelated between the processes, following the studies in Ref. [70].
The uncertainties in the theoretical cross-section calculations are process dependent and vary from 4% for the tchannel to 6% for tt (see Sect. 3). In addition, there are large uncertainties in the Z /W + jets production cross-sections. For every jet an additional uncertainty of 24% is assumed [71]. The uncertainty in the normalisation of W /Z -boson production in association with three jets is 42%, and the rate of W -boson events with heavy-flavour jets is allowed to vary by an additional 20%.
The modelling of the W + jets background was crosschecked using Alpgen with Pythia. The shape of the W+jets background was found to be consistent with the nominal Sherpa prediction. Hence no dedicated systematic uncertainty is assigned to the choice of generator, in order to avoid double-counting of the statistical uncertainty of the prediction (model statistics).
Uncertainties related to the modelling of the fake-lepton background take into account the choice of control region for the determination of the fake-and real-lepton efficiencies, the choice of parameterisation, and the normalisation of the prompt-lepton backgrounds in the determination of the efficiencies [52].
The uncertainty due to the limited size of the simulated samples and the fake-lepton background (model statistics) is estimated through the procedure detailed in Refs. [72,73]: for every bin of the discriminant, an independent parameter is assigned which describes the variation of the predicted event rate constrained by its statistical uncertainty.

Statistical analysis
A binned profile maximum-likelihood fit to the discriminant in the signal region is used to determine the tW cross-section. The likelihood function is defined as a product of Poisson probability terms over all the bins of the discriminant in the signal region and Gaussian penalty terms, where the n i (ν i ) is the observed (expected) number of events in each bin i of the discriminant. The expected number of events depends on the signal-strength parameter, μ, which is a multiplicative factor on the predicted signal cross-section. Nuisance parameters (NPs), θ k , are used to encode the effects of the systematic uncertainties in the expected number of events. The Gaussian penalty terms model the external constraints on these parameters. The estimated parameters, denoted byμ andθ , are obtained by maximising L(μ, θ ; n).
The likelihood function is composed and evaluated with the HistFactory program [74], part of the RooStats framework [75]. The minimisation is performed with the Minuit package [76], using Minos to compute the error estimates.
The statistical significance, Z , of the result is estimated by comparing the likelihood values of two hypotheses. The background-only hypothesis is that there is no signal in the data (or equivalently, μ = 0). The signal-plus-background hypothesis is that the signal exists with the signal strength obtained from the fit to data. With the asymptotic approximation [77], the significance is calculated using a test statistic based on the profile likelihood ratio, whereˆ θ μ=0 denotes the estimates of the nuisance parameters that maximise the likelihood function under the backgroundonly hypothesis. The expected significance is calculated by replacing n in the likelihood function with the Asimov dataset for the nominal signal-plus-background hypothesis (μ = 1, θ =ˆ θ).

Cross-section measurement
The tW cross-section is extracted from the fit to data in the signal region. Given the Standard Model prediction, the extracted signal strength is expected to beμ = 1.00 ± 0.35. The measured value isμ = 1.16 ± 0.31, corresponding to an observed cross-section of σ obs tW = 26 ± 7 pb, which is con- sistent with the Standard Model prediction. The observed (expected) significance is 4.5σ (3.9σ ). The (post-fit) impact of each systematic uncertainty on the measured signal strength is estimated by means of conditional fits, i.e. the fit is repeated while keeping the corresponding nuisance parameter fixed at the ±1 standard deviation (sigma) value of the post-fit error interval. The resulting change in the estimate of the signal strength quantifies the impact of the uncertainty. For each nuisance parameter, the +1 and −1 sigma variations are found to be symmetric about the best-fit value to a very good approximation. Table 2 shows the impacts of the systematic uncertainties on the observed fit result, where the impacts of uncertainties with similar sources have been added in quadrature. The dominant uncertainties are due to the amount of QCD radiation in signal events and tt background, the JES and b-tagging, and the model statistics, including the limited size of the MC samples.
Some nuisance parameters are constrained by the data. For example, the normalisation uncertainty for W +jets events is reduced from 45% to 8%, because the assigned initial uncertainty is large and this background can be separated well from tW and tt events. By design of the discriminant, combinations of nuisance parameters that shift the peak in the m(W H ) distribution are constrained, primarily the JES and choice of renormalisation scale together with the amount of Fake leptons 3000 ± 1600 5 ± 22 Total background 126900 ± 1900 38900 ± 2400 Total signal + background 134700 ± 500 40200 ± 2300 Observed 134633 41738 QCD radiation in signal and tt background. Also, the nuisance parameter for the NLO matching for tW and tt is constrained: the choice of MC@NLO is not supported by the data, reducing the impact of the choice from 9% pre-fit to 3% post-fit. A few nuisance parameters are pulled away from the pre-fit expectation. For the parameter associated with the choice of parton-shower generator, a blend of Pythia and Herwig gives the best description of the data, while the nominal Pythia prediction is disfavoured at the two-sigma level. The b-tagging parameter with the largest effect on the overall b-tagging efficiency is pulled by about one sigma, corresponding to a decrease of about 1% to 2% in the btagging efficiency compared to the pre-fit expectation. Given that the b-tagging calibration partially relies on dijet events [61], which correspond to a different environment regarding the production mechanism of the b-jets, the pull is reasonable. Table 3 shows the post-fit event yields of each process. The uncertainties in the yields are computed taking the correlations between nuisance parameters and processes into account. The post-fit estimates are well within the uncertainties of the pre-fit expectation (Table 1), while most of their uncertainties are reduced. The normalisation uncertainty for W + HF jets changes from almost 50% to about 10%. Figure 7 shows the post-fit distributions for the NN input variables, the NN output response and the m(W H ) in the signal region. The post-fit plots use the parameter estimates obtained in the fit of the discriminant, including their uncertainties, and demonstrate a good description of the data. Figure 8a shows that the data are well described by the model in the signal region. Figure 8b shows the strongest support for the validity of the fit result by comparing the expected distributions and observed distributions in the tt validation region. It shows that the uncertainty due to the extrapolation from the signal region is small, and therefore provides a stringent test that the main background is well modelled.

Conclusion
The inclusive cross-section for the production of a single top quark in association with a W boson in the singlelepton channel is measured using an integrated luminosity  [78].

Data Availability Statement
This manuscript has no associated data or the data will not be deposited. [Authors' comment: "All ATLAS scientific output is published in journals, and preliminary results are made available in Conference Notes. All are openly available, without restriction on use by external parties beyond copyright law and the standard conditions agreed by CERN. Data associated with journal publications are also made available: tables and data from plots (e.g. cross section values, likelihood profiles, selection efficiencies, cross section limits, ...) are stored in appropriate repositories such as HEPDATA (http:// hepdata.cedar.ac.uk/). ATLAS also strives to make additional material related to the paper available that allows a reinterpretation of the data in the context of new theoretical models. For example, an extended encapsulation of the analysis is often provided for measurements in the framework of RIVET (http://rivet.hepforge.org/)." This information is taken from the ATLAS Data Access Policy, which is a public document that can be downloaded from http://opendata.cern.ch/record/413 [opendata.cern.ch].] Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecomm ons.org/licenses/by/4.0/. Funded by SCOAP 3 .