Review of single vector boson production in pp collisions at $\sqrt{s} = 7$ TeV

This review summarises the main results on the production of single vector bosons in the Standard Model, both inclusively and in association with light and heavy flavour jets, at the Large Hadron Collider in proton-proton collisions at a center-of-mass energy of 7 TeV. The general purpose detectors at this collider, ATLAS and CMS, each recorded an integrated luminosity of $\approx 40\,{\rm pb^{-1}}$ and $5\,{\rm fb^{-1}}$ in the years 2010 and 2011, respectively. The corresponding data offer the unique possibility to precisely study the properties of the production of heavy vector bosons in a new energy regime. The accurate understanding of the Standard Model is not only crucial for searches of unknown particles and phenomena but also to test predictions of perturbative Quantum-Chromo-Dynamics calculations and for precision measurements of observables in the electroweak sector. Results from a variety of measurements in which single W or Z bosons are identified are reviewed. Special emphasis in this review is given to interpretations of the experimental results in the context of state-of-the-art predictions.


Introduction
The W and Z bosons, since their discovery at UA1 [1,2] and UA2 [3,4] in the early 1980s, have been the subject of detailed measurements at both electron-positron and hadron colliders. The ALEPH, DELPHI, L3 and OPAL experiments at the large electron-positron collider, LEP, preformed many precision studies of these vector bosons, including measurements of the branching ratios [5], the magnetic dipole moment and the electric quadrupole moment [6], all of which were measured with a precision of better than 1%. At hadron colliders, single vector boson production has been explored at √ s = 0.63 TeV at the CERN SppS by UA1 and UA2, and at both √ s = 1.8 TeV and √ s = 1.96 TeV at the Tevatron by CDF [7,8] and D0 [9,10]. The distinct advantage of W and Z production measurements at the hadron colliders is that the number of single vector boson events is large, roughly 138,000 Z → ee and 470,000 W → eν candidates using 2.2 fb −1 of data at CDF [11,12] with relatively low background rates, roughly 0.5% of Z candidates and 1% of W candidates at CDF. The major disadvantage is that the parton center-ofmass energy can not be determined for each event because of the uncertainties in the structure of the proton. Despite these challenges, the CDF and D0 experiments have reported measurements of the mass of the W [12,13] with a precision comparable to the measurements at LEP. In addition with the large data samples, measurements of W and Z production in association with jets [14][15][16][17][18][19] and the production of W and Z production in association with heavy flavour quarks [20][21][22][23][24][25] were preformed. Measurements of single vector boson production at the SppS and the Tevatron have been especially important for the development of leading-order and next-to-leading order theoretical predictions, most of which are used today for comparisons to data at the Large Hadron Collider (LHC). Finally W and Z production has also been observed at heavy-ion colliders at RHIC at √ s = 0.5 TeV [26,27] and the LHC at √ s = 2.76 TeV [28][29][30]. For a detailed review of the measurements preformed at LEP and SppS see [31] and [32] respectively.
Today, the focus of measurements of W and Z production at the LHC is to test the theory of perturbative Quantum-Chromodynamics (QCD) in a new energy regime, to provide better constraints on the parton distribution functions, and to improve electroweak precision measurements, such as the mass of the W and sin 2 θ . As W and Z production are dominant backgrounds to Higgs boson measurements and searches for physics beyond the Standard Model, these new measurements also provide insight to these studies.
For tests of the predictions of perturbative QCD, the benefit of the increase in energy at the LHC can readily be seen in Figure 1, where the Bjorken x-values of the interacting partons for a given process, e.g. the production of a Z boson, is shown. The reach in the low-x region has been increased by more than two orders of magnitude compared to that of the SPS and the Tevatron. As a matter of fact, these new measurements not only benefit from the higher center-of-mass energy but also from improved statistical and systematic uncertainties. At the LHC, copious amounts of W and Z boson events, more than a million Z → ee events at each of the ATLAS and CMS experiments during the 2011 √ s = 7 TeV run, were detected, with an improved experimental precision. For example, the uncertainty on the jet energy scale is almost a factor of three better [33,34] compared to that at the Tevatron experiments. Furthermore, the detectors have been designed to have an increased rapidity acceptance and can measure electrons for some measurements to |η| < 4.9 and jets to |η| < 4.4. As a result, a large fraction of these low-x events shown in Figure 1 can be reconstructed by the LHC detectors.
The theoretical predictions used for comparison to these measurements have been extended and improved. For inclusive W and Z production, theoretical predictions at next-tonext-to-leading order in perturbation theory are available. For measurements of single vector boson production in association with jets predictions at next-to-leading order for up to five additional partons in the final state exist. The magnitude of the theoretical uncertainties in these calculations are comparable to those of the experimental uncertainties. In addition several advanced leading-order predictions exist which simulate the entire event process from the hard scatter to the parton showering and the fragmentation. Although many of these predictions have been vetted by measurements at previous hadron colliders, the LHC measurements will test these predictions in previously unexplored regions of the phase space.
The structure functions of the proton, which are a dominant source of uncertainties in electroweak precision measurements at hadron colliders, can also be constrained through studies of the differential cross-sections of W or Z bosons production. This is illustrated in Figure 2, where the kinematic plane as a function of the Bjorken x and Q 2 for Drell-Yan scattering at the Tevatron, the LHC and the corresponding deep inelastic scattering experiments are compared. Similarly, measurements of the W production in association with a charm quark test the contributions in the proton from the strange quarks at x ≈ 0.01 as well as any s −s asymmetries. This review article summarises the major results of the single W and Z production in the proton-proton collision data at √ s = 7 TeV recorded in the years 2010 and 2011 at the LHC by the two general purpose experiments, AT-LAS and CMS. The article is organised as follows. First, in Section 2, we review the basic theory behind single vector boson production. In this section, we pay special attention to some of the basic elements of cross-section calculations such as the matrix element calculations, the parton shower modelling and parton distribution functions. We also summarise here the theoretical predictions used in this review. In Section 3, we describe the ATLAS and CMS detectors at the LHC and discuss in a general manner the basic principles of lepton and jet reconstruction. Section 4 delineates how cross-sections are measured at the LHC, while Section 5 highlights the event selection and the background estimates for the measurements presented here. Finally we present the results for inclusive single vector boson production in Section 6 as well as the production in association with jets in Section 7. In Section 8, we conclude and provide thoughts on future measurements.

Vector Boson Production in the Standard Model
The electroweak Lagrangian of the Standard Model after electroweak symmetry breaking, i.e. after the Higgs Boson has acquired a vacuum expectation value, can be written as [35] where the terms in Equation 1 are schematically illustrated as tree level Feynman graphs in Figure 3. The kinetic term, L K , describes the free movement all fermions and bosons. It involves quadratic terms and the respective masses. The term L N describes the interaction of the photon and the Z boson to fermions, while L C describes the interaction of the W-Boson to left-handed particles or right-handed antiparticles. The self interaction of gauge bosons is a direct consequence of the SU(2) L group structure and is described by L WWV and L WWVV , representing three-point and fourpoint interactions, respectively. The three-and four-point self-interaction of the Higgs boson is described by L H , while the interaction of the Higgs boson to the gauge bosons is represented in L HV . The last term in Equation 1, L Y characterises the Yukawa couplings between the massive fermions of the Standard Model and the Higgs field. For the single Z boson production at hadron colliders, L N is where J em µ describes the electromagnetic current, i.e. the sum over all fermion fields weighted by their electromagnetic charges, and J 3 µ represents the weak current, involving only left-handed particles and right-handed anti-particles with their f f " + " " +" " + " " + " " +" " + " " + " respective weak isospins. The weak mixing angle, θ W , describes the relative contribution of the weak and electromagnetic part of the interaction. The production of the single W bosons is described by where only the terms for first generation are explicitly shown. The quark and lepton spinor fields are denoted by u i , d j and v i , e i . The term (1−γ 5 ) acts as a projector for the left-handed components of the spinors, meaning that the charge current acts exclusively on left-handed particles and right-handed anti-particles, while for the neutral current all spinor components play a role due to the electromagnetic part of the interaction term. The Cabbibo-Kobayashi-Maskawa (CKM) matrix is denoted as V CKM i j [36]. In this review article, we will concentrate on the terms of the Electroweak Lagrangian represented in Equation 2 and 3, as they describe the single vector boson production within the Standard Model.
Before discussing the experimental results, we first review the central parts of the theoretical predictions of the gauge boson production cross-sections at the LHC. In Section 2.1 the calculation of the cross-section is defined, which is shown to consist of two main parts; the matrix-element term describing the parton interactions and the parton distribution functions describing the proton. As the lowest-order matrix-element term for W and Z production is a frequent example in many particle physics textbooks, we extend the formalism in Section 2.2 by discussing higher-order corrections from the QCD and Electroweak theories and emphasise why these are important to the experimental measurements. The hard scatter process, which at high energy scales can be connected to the lower energies scales via parton showering models is discussed in Section 2.3. Finally, the second part of the cross-section calculation, the parton distribution functions, is briefly reviewed in Section 2.4. In addition, we discuss some critical inputs which are needed to perform these calculations. This includes the available models for hadronisation of final state particles (section 2.5) and the description of multiple particle interactions (section 2.6). A discussion of the available computing codes, which are used to compare the latest LHC measurements with the predictions of the Standard Model, can be found in Section 2.7. Our discussion ends with an overview of the definition and interpretation of some observables which are important for understanding QCD dynamics (section 2.8).
Several introductory articles on the production of vector bosons in hadron collisions are available. We summarise here the essential aspects along the lines of [37], [38]. Related overview articles on parton density functions at nextto-leading order at hadron colliders and subsequently jet physics in electron-proton can be found in [39] and [40], respectively.

Cross-section calculations
The calculation of production cross-sections in proton-proton collisions at the LHC is, in general, a combination of two energy regimes: the short-distance or high-energy regime and the long-distance or low-energy regime. By the factorisation theorem, the production cross-section can therefore be expressed as a product of two terms: one describing the partonparton cross-section,σ qq→n at short-distances and another describing the complicated internal structure of protons at long distances. For large momentum transfers of the interacting partons in the short-distance term, the parton-parton interaction can be evaluated using perturbative QCD calculations. However, in the long-distance term, where perturbative calculations are no longer applicable, parton density functions (PDFs) are used to describe the proton structure in a phenomenological way. These functions are written as f a/A (x, Q 2 ) for the parton a in the proton A where x = p a p A is the relative momentum of the parton in direction of the proton's momentum and Q 2 is this energy scale of the scattering process. The scale at which the long-distance physics of the PDF description and the short-distance physics of partonparton interaction separate is called the factorisation scale and is defined as µ F = Q. For the production of a vector boson via quark-fusion, the energy scale is set to the mass of the vector boson, which in turn can be expressed by the invariant mass of final state fermions f , i.e. Q 2 = m 2 V = m ff . The proton-proton cross-section is thereby expressed as and shown graphically for Z production in Figure 4. The functions f a/A and f b/B denote the PDFs for the partons a and b in protons A and B. All quark flavours are accounted for in the sum and the integration is performed over x a and x b , describing the respective momentum fractions of the interacting partons. The subset of perturbative corrections from real and virtual gluon emissions, which are emitted collinearly to the incoming partons, lead to large logarithms that can be absorbed in the PDFs.
Inclusive hard-scattering processes can be described using the factorisation theorem [41], [42]. This approach is also applicable when including the higher order perturbative QCD corrections, which are discussed in more detail in Section 2.2.2. When expanding the parton-parton cross-section in terms of α s , the formula for the cross-section becomes: where σ 0 is the tree-level parton-parton cross-section and σ 1 is the first order QCD correction to the parton-parton crosssection, etc. The renormalisation scale, µ R is the reference scale for the running of α s (µ 2 R ), caused by ultraviolet divergences in finite-order calculations.
Writing this equation in terms of the matrix elements yields where 1/(2ŝ) is the parton flux, φ n is the phase space of the final state and |m qq→n | is the corresponding matrix element for a final state n, which is produced via the initial state qq. The matrix element can then be evaluated according to perturbation theory as a sum of Feynman diagrams, qqn . The evaluation of these integrals over the full phase space is typically achieved via Monte Carlo sampling methods.

Leading-order calculations
The calculation of the electron-positron annihilation crosssection in pure Quantum Electrodynamics (QED), i.e. e + e − → µ + µ − , is straightforward and can easily be extended to the quark-antiquark annihilation cross-section by including the colour factor of 1/3 and accounting for the charge Q q of the involved quarks q: whereŝ = (x A P A +x B P B ) 2 = x A x B s and √ s denotes the centerof-mass energy of the proton-proton collision. In an electroweak unified theory, the cross-section must also include the exchange of a Z boson for larger energies ( √ s > 40 GeV) and is therefore extended bŷ The vector and axial couplings of the Z bosons to the leptons and quarks are denoted as g V = g L + g R and g A = g L − g R which can be expressed as combinations of left-and righthanded chiral states for the quarks q and leptons l. The Z boson propagator K(ŝ) can be written as In the narrow-width approximation, the Z boson is assumed to be a stable particle and the propagator reduces to a δfunction. This approximation is based on the fact that the width of the Z boson ( Γ Z ≈ 2.5 GeV) is small compared to its mass (m Z ≈ 91 GeV). Hence the parton-parton crosssection can be expressed aŝ when omitting the interference with the photon-exchange in the s-channel 1 .
The decay of the Z boson into fermion pairs is described by the branching ratio Br(Z → ff ) = Γ (Z → ff )/Γ Z , where the partial width Γ (Z → ff ) is given in lowest order by where the factors g f V and g f A are again the vector-and axial couplings for the respective fermions f to the Z boson. The colour factor N C is 1 for leptons and 3 for quarks. This leads to a prediction of ≈ 70% decays into quark and antiquarks, but only ≈ 3.4% for the decay into a single generation of charged leptons.
The lowest-order cross-section for the W boson production via quark-antiquark fusion can be derived in a similar manner. In contrast to the Z boson production, quarks from different generations can couple to the W boson, while the interference with the electromagnetic sector is not present. The cross-section in the narrow width approximation is given bŷ where the CKM matrix element accounts for the quark-generation mixing. The partial decay width of the W boson at lowest order is leading to 1/3 probability for leptonic decays and 2/3 for decays into quark/antiquark pairs.

Perturbative QCD corrections and jets
The leading-order calculations of the W and Z boson production, as shown in Section 2.2.1, suggest that the momentum of the boson in the transverse plane is zero. Yet it is well known from collider experiments that the transverse momentum (p T ) of W and Z bosons peaks at few GeV, with a pronounced tail to high values, i.e. p T m V [43], [44], [45], [46]. To understand the physical origin of this, two different effects have to be taken into account. First, the interacting partons are believed to have an intrinsic transverse momentum (k T ) relative to the direction of the proton, leading to an exponentially decreasing p T distribution of the vector boson.
The experimentally determined value of the average intrinsic momentum is < k T >= 0.76 GeV, measured in protonneutron collisions [47] and is not large enough to explain the observed p T distribution of vector bosons in hadron collisions. The second, larger effect arises from higher order QCD corrections to the vector boson production, which can lead to the radiation of additional quarks and gluons in the final state in the transverse plane. The vector sum of these emissions has to be balanced by the transverse momentum of the produced vector boson, which in turn acquires a transverse momentum. In the regime where α s is small, these perturbative QCD corrections may be calculated. The two contributing classes of next-to-leading-order (NLO) corrections, i.e. the virtual loop corrections and the real emissions of gluons/quarks, are illustrated in Figure 5. The correction terms with virtual loops do not affect the transverse momentum spectrum of the vector boson directly. The real corrections however, imply the existence of 2 → 2 processes, leading to an additional parton in the final state which boosts the W or Z boson in the transverse plane. The generic form of the production cross-section for the processes qq → V g and qg → V q , where V stands for a vector boson, can be expressed by the Madelstam variables, describing the Lorentz invariant kinematics of a 2 → 2 scattering process. The resulting cross-section at NLO is proportional to As t, u → 0, divergencies in Equation 13 occur. This can be interpreted as final state quarks or gluons with a vanishing transverse momentum, i.e. those which are collinear to the incoming parton. Therefore, a minimal p T requirement of the additional quark or gluon in the final state needs to be applied to obtain a finite production cross-section prediction. In the calculation of the fully inclusive production cross-section, the divergencies are compensated by the virtual loop corrections. Two main energy regimes of the transverse momentum spectrum of the vector boson production are considered here: A high energy regime, where p T m V and an intermediate The linear dependence of α s is a consequence of the NLO QCD corrections, leading to one additional parton in the final state (V+1 jet production). Each additional parton in the final state requires one additional higher order QCD correction and therefore an additional order of α s . Examples of leading-order Feynman diagrams for the Z+2 jet production are shown in Figure 6. For QCD corrections with multiple jets, the probability that an additional parton is a radiated gluon is governed by a Poisson distribution. This implies that the leading-order term for a V+n-jet final state, called Poisson scaling, has the form of where σ tot is the total cross-section,n is a expectation value of the Poisson. This is the expected behaviour at e + e − colliders [48], where PDFs do not play a role. However, at hadron colliders, the experimentally observed V+n-jet final state, called staircase scaling, has the form of where the coefficients a depend on the exact definition of a jet and σ 0 is the zero-jet exclusive cross-section. The ratio of the n-jet and (n+1)-jet cross-sections is then a constant value, σ n+1 σ n = e −a , where e −a is a phenomenological parameter which is measurement dependent. The reason for observed staircase scaling at hadron colliders is two-fold. At small numbers of additional partons, the emission of an additional parton is suppressed in the parton density function. At larger numbers of additional partons, the probability of gluon radiation no longer follows a Poisson distribution due to the non-abelian nature of QCD theory, which states that a gluon can radiate from another gluon. For large jet multiplicities a deviation from the staircase scaling behaviour is expected, as the available phase space for each additional jet in the final state decreases.
Today, several leading-order calculations, such as [49] and [50], are available that describe more than six partons in the final state. The inclusive production cross-section and the associated rapidity distribution for vector bosons is known today to next-to-next-to-leading order (NNLO) [51], [52], [53].
The intermediate momentum range of k T < p T (V ) < m V /2 can also be assessed with perturbative calculations. However, higher-order corrections, manifested as low energetic gluons emitted off the incoming partons at intermediate energies, must be included for a correct description of the experimental data. This can be most easily seen in the limit of t → 0 and u → 0; then the final state gluon in the qq → V g process becomes collinear to the incoming parton. The corresponding Feynman diagram can be redrawn as initial state radiation (ISR) as shown in Figure 7.  The main contributions of these collinear gluon emissions to the cross-section at the n-th order are given by Such collinear gluon emissions are also the basis for parton showers, which will be discussed in Section 2.3. Summing up the gluon emissions to all orders leads to where C F = 4/3 is the QCD colour factor for gluons. This approach, known as Resummation [54], has been significantly improved and extended in recent years and provides currently the most precise predictions for the transverse momentum distribution of vector bosons in the low energy regime.

Electroweak corrections
So far, only QCD corrections to W and Z boson production have been discussed. The virtual one loop QED corrections and the real photon radiation corrections are illustrated via Feynman diagrams in Figure 8. The NLO corrections to the charged and neutral currents are well known [55], [56]. In particular, the full O(α em ) corrections to the pp → γ/Z → l + l − process with O(g 4 m 2 T /m 2 W ) corrections to the effective mixing angle sin 2 θ 2 e f f and m W are available [57][58][59][60][61]. Even though the electroweak corrections to the vector boson production cross-section are small compared to the higher-order QCD corrections, they lead to a significant distortion of the line shape of the invariant mass and subsequently to the transverse momentum spectrum of the decay leptons. The comparison between the corresponding distributions with and without electroweak corrections is illustrated in Figure 9.   Fig. 9 Electroweak (EWK) corrections to the lepton transverse momentum for the neutral current Drell-Yan process at the LHC [62]. Results are presented for bare electrons and electrons employing electronphoton recombination, respectively.
In general, electroweak corrections at moderate energies are dominated by final-state radiations (FSR) of photons, which is indicated by the upper right diagram in Figure 8. Certain measurements, like the determination of the W boson mass from the decay lepton p T distribution, are sensitive to these corrections. For this measurement, this effect can induce a shift of up to 10 MeV on m W in the muon decay channel. In contrast, the electron decay channel is less effected due to the nature of electron reconstruction in the detector, where the FSR photons are usually reconstructed together with the decay electron. Therefore the relative magnitude of the electroweak corrections with respect to the QCD corrections must be considered individually for each measurement. It should be noted that there are also diagrams with photons in the initial state and hence the parton density functions evolve with combined QCD and QED evolution equations. That can lead to sizable corrections in the high and low mass Drell-Yan production [63].

Parton shower models
As discussed in the previous section, matrix-element calculations at fixed-order provide cross-sections assuming that the final state partons have large momenta and are well separated from each other. Parton shower models provide a relation between the partons from the hard interaction (Q 2 Λ QCD ) to partons near the energy scale of Λ QCD . Here Λ QCD is defined as the transition energy between the high-energy and low-energy regions. A commonly used approach for parton shower models is the leading-log approximation, where showers are modelled as a sequence of splittings of a parton a to two partons b and c. QCD theory allows for three types of possible branchings, q → qg, g → gg and g → qq, while only two branchings exist in QED theory, namely q → qγ and l → lγ.
The differential probability dP a for a branching for QCD emissions is given by the Dokshitzer-Gribov-Lipatov-Altarelli-Parisi (DGLAP) evolution equations [64][65][66] with the evolution time t defined as t = ln(Q 2 /Λ 2 QCD ). Q 2 denotes the momentum scale at the branching and z the momentum fraction of the parton b compared to parton a. The sum runs over all possible branchings and P a→b,c denotes the corresponding DGLAP splitting kernel.
This relation also holds for the QED branchings, where the coupling constant in Equation 19 is replaced with α em . For final state radiation, the emission of particles due to subsequent branchings of a mother parton is evolved from t = Q 2 hard at the hard interaction to the non-perturbative regime t ≈ Λ QCD . Initial-state radiation can be ordered by an increasing time, i.e. going from a low energy scale to the hard interaction. This can be interpreted as a probabilistic evolution process connection two different scales: the initial scale Q 2 0 of the interaction to the scale of the hard interaction scale Q 2 hard . During this evolution, all possible configurations of branchings, leading to a defined set of partons taking part in the hard interaction, are considered.
The implementation of parton showers is achieved with Monte Carlo techniques. They are used to calculate the steplength t 0 to t where the virtuality of the parton decreases with no emissions. At the evolution time t, a branching into two partons occurs where the resulting sub-partons have smaller virtuality than the initial parton. This procedure is then repeated for the sub-partons, starting at the new evolution time t 0 = t.
Therefore the probability for the parton a at the scale t 0 not to have branched when it is found at the scale t has to be determined. This probability P no−branching (t 0 ,t) is given by the Sudakov form factor [67], which can be expressed as where is the differential branching probability for a given evolution time t with respect to the differential range dt . The latter relation follows directly from Equation 19, by integrating over the allowed momentum distributions z. The probability for a branching of a given parton a at scale t can then be expressed by the derivative of the Sudakov form factor S a (t): This relation describes the effect known as Sudakov suppression: The first factor I a→b,c , which describes the branching probability at a given time t, is suppressed by the Sudakov form factor S a (t), i.e. by taking into account the possibility of branchings before reaching the actual scale t.
The branching of the initial-state partons during the parton shower therefore leads to the emission of gluons or quarks, which in turn may add an additional jet to the event. The final-state partons predicted within the leading-log approximation are dominated by soft and collinear radiations and hence large momentum jets that are not expected to be described correctly within this approximation. The kinematics of the missing hard scatter components are predicted by the corresponding higher-order diagrams, which have been discussed in Section 2.2.2.
The main advantage of the parton shower approach is its simplicity compared to matrix-element calculations which increase in complexity when considering more independent partons in the initial and final states. However, there is an important difference between the soft-gluon emission described by parton shower and the emission of a gluon calculated by NLO matrix element. While the full matrix-element calculation includes the spin-1 nature of the gluon and hence induced polarisation effects on the intermediate gauge boson, the parton shower algorithm does not take into account spin effects.

Parton distribution functions and scale dependencies
The PDFs play a central role not only in the calculation of the cross-section in Equation 5, but also in the modelling of parton showers and hadronisation effects. A generic PDF f i (x, µ F , µ R ) describes at lowest order the probability of finding a parton of type i with a momentum fraction x when a proton is probed at the scale µ F . The factorisation and renormalisation scale parameters µ F and µ R in the PDF definition act as cut-off parameters to prohibit infrared and ultraviolet divergences. If a cross-section could be calculated to all orders in perturbation theory, the calculation would be independent from the choice of the scale parameters, since the additional terms would lead to an exact cancellation of the explicit scale dependence of the coupling constant and the PDFs. Both scales are usually chosen to be on the order of the typical momentum scales of the hard scattering process, to avoid terms with large logarithms appearing in the perturbative corrections. For the Drell-Yan process at leading order this implies µ F = µ R = m Z . It should be noted that both scales are usually assumed to be equal, even though there is no reason from first principles for this choice. The dependence of the predicted cross-section on µ F and µ R is thus a consequence of the missing/unknown higher-order corrections. The dependence is therefore reduced when including higher orders in the perturbation series. The uncertainty on the cross-section prediction due to scale uncertainties is usually estimated by varying both scales simultaneously within 0.5 · Q < µ F , µ R < 2 · Q, where Q is the typical momentum scale of the hard process studied. However, this evaluation procedure sometimes provides results that are too optimistic and the differences between the leading-order and NLO calculations are not always covered by the above procedure.
As the actual form of f i (x, µ F ) cannot be predicted with perturbative QCD theory, a parameterised functional form has to be fitted to experimental data. The available data for the PDF determination comes mainly from deep inelastic scattering experiments at HERA, neutrino data, as well as Drell-Yan and jet production at the Tevatron and LHC colliders. Note that the scale dependence of f i is predicted by the DGLAP evolution equations.
In order to fit PDFs to data, a starting scale, where perturbative QCD predictions can be made, has to be chosen and a certain functional form of the PDFs has the be assumed. A typical parametrisation of f i (x, µ F ) takes the form where P is a polynomial function and a j are free fit parameters which cannot be predicted from perturbative QCD calculations, but can only be determined by experiment. In a second step, a factorisation scheme, i.e. a model for the handling of heavy quarks, and an order of perturbation theory has to be chosen. The DGLAP evolution equations can then be solved in order to evolve the chosen PDF parametrisation to the scale of the data. The measured observables can then be computed and fitted to the data. The PDF fits are currently performed and published for leading-order, NLO and NNLO calculations. Even though most matrix elements are known to NLO order in QCD theory, some parton shower models are still based on leading-order considerations and therefore leading-order PDF sets are still widely used. PDF fitting is performed by several groups. The CTEQ-TEA [68], MSTW [69], ABKM [70], GJR [71] and NNPDF [72] collaborations include all available data for their fits, but face the problem of possible incompatibilities of the input data, caused by inconsistent measurement results from different experiments. These results differ in the treatment of the parametrisation assumptions of f i . The HERAPDF [73] group bases their PDF fits on a subset of the available data, where only the HERA measurements have been chosen as input and therefore possible inconsistencies in the fit are reduced.
It should be highlighted that the PDF approach and fitting is subject to several assumptions and model uncertainties. The actual form of the input distributions is arbitrary and hence the choice of the analytical function implies a model uncertainty. The approach of the NNPDF group is an exception as the parametrisation is chosen to be handled by a flexible neural network approach. In addition it is commonly assumed that the strange-quark content follows s =s = (ū +d)/4. The suppression of s ands-quark content is due to their larger masses compared to theū and d quarks, but a rigorous argument of the chosen suppression factor cannot be derived from first principles. Similar is the situation for heavy-flavour (c, b,t-quarks) contributions to the proton structure. Their contribution is 0 below the Q 2threshold and is evolved according to the DGLAP-equations above.
The results presented in this paper rely mainly on the CTEQ-10 and MRST PDF sets [68], [69]. The PDF set for     Figure 11, where the x-values of both interacting quarks per events have been used.
The associated uncertainties of a given PDF set are based on the Hessian method [74], where a diagonal error matrix with corresponding PDF eigenvectors is calculated. The matrix and the PDF eigenvectors are functions of the fitparameters a i . Non-symmetric dependencies are accounted for by using two PDF errors for each eigenvector in each direction. The PDF uncertainties on a cross-section is then given by where a ± i labels the corresponding eigenvectors of a chosen PDF error set. This approach assesses only the PDF fit uncertainties within a certain framework, i.e. for a chosen parametrisation and various base assumption. Hence, usually also the difference between two different PDF sets from two independent groups, e.g. CTEQ and MRST, are taken as an additional uncertainty. The same procedure for the impact of PDF related uncertainties can be applied for any observable and is not restricted to inclusive cross-sections.

Hadronisation
The process of how hadrons are formed from the final state partons is call hadronisation. The scale at which the hadronisation is modelled is Q 2 = Λ 2 QCD . Since this process is complex, phenomenological models must be used. A detailed discussion can be found elsewhere [75].
The first models of hadronisation were proposed in the 70s [76] and today, two models are widely in use. The socalled string model [77] is based on the assumption of a linear confinement and provides a direct transformation from a parton to hadron description. It accurately predicts the energy and momentum distributions of primary produced hadrons, but requires numerous parameters for the description of flavour properties, which have been tuned using data. The second approach for the description of hadronisation is known as cluster model [78], [79], which is based on the pre-confinement property of parton showers [80]. It involves an additional step before the actual hadronisation, where colour-singlet subsystems of partons (denoted as clusters) are formed. The mass spectrum of these clusters depend only on a given scale Q 0 and Λ QCD , but not on a starting scale Q, with Q Q 0 > Λ QCD . The cluster model has therefore fewer parameters than the string model, however, the description of data is in general less accurate.
The subsequent decay of primary hadrons is either directly implemented in the computing codes for hadronisation, or in more sophisticated libraries such as EVTGEN [81]. Special software libraries can be used for the description of the τ-lepton decay, e.g. TAUOLA [82], correctly taking into account all branching ratios and spin correlations. Since hadronisation effects usually have only a small impact on the relevant observables discussed in this article, we refer to [75] for a detailed discussion.

Multiple-parton interactions
Equation 5 describes only a single parton-parton interaction within a proton-proton collision. However, in reality, several parton-parton interactions can occur within the same collision event. This phenomenon is known as multiple-parton interactions (MPI). Most of the MPI lead to soft additional jets in the event which cannot be reconstructed in the detector due to their small energies. Hence they contribute only as additional energy deposits in the calorimeters. However, a hard perturbative tail of the MPI, following ∼ d p 2 T /p 4 T , where p T is the transverse momentum of the additional jets, can lead to additional jets in the experimental data. These effects must be taken into account for the study of vector boson production in association with jets. Further information about the current available models for MPI can be found in [83]. Dedicated studies of MPI have been done at the LHC using W events with two associated jets [84]. The fraction of events arising from MPI is 0.08 ± 0.01(stat.) ± 0.02(sys.) for jets with a p T > 20 GeV and a rapidity |y| < 2.5. This fraction decreases when the p T requirements on the jets increases.

Multiple purpose event generators
Multiple purpose event generators include all aspects of the proton-proton collisions: the description of the proton via an interface to PDF sets, initial-state shower models, the hard scattering process and the subsequent resonance decays, the simulation of final-state showering, MPI, the hadronisation modelling and further particle decays. Some frequently used generator in the following analyses are PYTHIA6 [85], PYTHIA8 [86], HERWIG [87], HERWIG++ [88] and SHERPA [89]. All of these generators contain an extensive list of Standard Model and new physics processes, calculated with fixed-order treelevel matrix elements. Some important processes, such as the vector boson production are also matched to NLO crosssections.
The PYTHIA generator family is a long established set of multiple purpose event generators. While PYTHIA6, developed in Fortran in 1982 is still in use, the new PYTHIA8 generator was coded afresh in C++. The showering in PYTHIA6 used in this review is implemented with a p T -ordered showering scheme, whereas the new version used here is based on a dipole showering approach. The hadronisation modelling in both versions is realised via the Lund string model. MPI are internally simulated in addition.
Similar to PYTHIA, the HERWIG generator was originally developed in Fortan and is now superseded by HER-WIG++, written in C++. Both versions use an angular-ordered parton shower model and the cluster model for hadronisation. The JIMMY library [90] is used for the simulation of MPI.
The SHERPA generator was developed in C++ from the beginning and uses the dipole approach for the parton showering. The hadronisation is realised with the cluster model. The MPI are described with a model that is similar to the one used in PYTHIA. For the analyses described here, SHERPA is generated with up to five additional partons in the final state.

Leading-order and NLO matrix-element calculations
Several programs such as ALPGEN [49] and MADGRAPH [91] calculate matrix elements for leading-order and some NLO processes, but do not provide a full event generation including parton shower or hadronisation modelling. These generators as well as SHERPA are important because they contain matrix-element calculations for the production of vector bosons in association with additional partons. ALPGEN is a leading-order matrix-element generator and includes predictions up to six additional partons in the final state. This is achieved by adding real emissions to the leading-order diagrams before the parton shower modelling. In this way, although the process is calculated at leading-order, tree-level diagrams corresponding to higher jet multiplicities can be included. Some of the virtual corrections are then added when a parton shower model is used. MADGRAPH for the analyses presented here follows a similar method and produces predictions up to four additional partons. The subsequent event generation, starting from the final parton configuration, is then performed by PYTHIA or HERWIG for ALP-GEN and PYTHIA for MADGRAPH.

Parton shower matching
There is significant overlap between the phase space of NLO or n-parton final-state QCD matrix-element calculations and the application to parton showers with respect to their initialand final-state partons, as both lead to associated jets. To avoid a potential double counting, matching schemes have been developed that allow matrix-element calculations for different parton multiplicities in the initial state and final state to be combined with parton shower models. The main strategies are based on re-weighting methods and veto-algorithms. The Catani-Krauss-Kuler-Webber (CKKW) matching scheme [92], [93] and the Mangano (MLM) scheme [94] are widely used for tree-level generators. For example, the ALPGEN generator uses the MLM scheme, whereas SHERPA uses CKKW matching for leading-order matrix-element calculation. A detailed discussion can be found in the references given.
An alternative, less generalised approach to matching schemes are merging strategies. Here the parton showers are reweighted by weights calculated by matrix-element calculations. In the PYTHIA generator only the first branching is corrected, while HERWIG modifies all emissions which could be in principle the hardest. These approaches model correctly one additional jet, but fail for higher jet multiplicities.

NLO generators
While matrix-element calculations give both a good description for the hard emission of jets in the final states and handle inferences of initial and final states correctly, they are not NLO calculations. A combined NLO calculation with parton shower models therefore is much desired. However, the above described methods work only for leading-order matrix-element calculations. For the matching between NLO matrix element and parton shower models more sophisticated methods have to be used. The MC@NLO approach [95] was the first available prescriptions to match NLO QCD matrix elements to the parton shower framework of the HER-WIG generator. The basic idea is to remove all terms of the NLO matrix-element expression which are generated by the subsequent parton shower. Due to this removal approach, negative event weights occur during the event generation. The aMC@NLO generator follows a similar approach for NLO calculations. The second approach is the POWHEG procedure [96], which is currently implemented in the POWHEG-BOX framework [97]. This framework allows for an automated matching of a generic NLO matrix element to the parton shower. The POWHEG procedure foresees that the hardest emission is generated first. The subsequent softer radiations are passed to the showering generator. In contrast to the MC@NLO approach, only positive weights appear and in addition the procedure can be interfaced to other event generators apart from HERWIG. PYTHIA8 also includes possibilities to match to NLO matrix element using the POWHEG scheme.

NLO calculations and non-perturbative corrections
MCFM [98] and BLACKHAT-SHERPA [99] provide NLO calculations up to two and five additional partons respectively. These calculations differ from NLO generators as they do not provide any modelling of the parton shower. These calculations compute both the virtual and real emission corrections for higher jet multiplicities. For the virtual corrections, the calculation is achieved by evaluating one-loop corrections to the tree-level diagrams, while the real emission corrections are obtained by matrix-element calculations which include an additional emitted parton. Several different techniques are used for these calculations, see for example [99].
Since MCFM and BLACKHAT-SHERPA are not matched to a parton shower model, they can not be directly compared to data or simulations which have parton showering and hadronisation applied to the final-state particles. To mimic the effects of both the parton shower and the hadronisation, non-perturbative corrections are estimated using a multiple purpose generator such as PYTHIA. These corrections are derived by comparing PYTHIA with and without the parton shower and hadronisation models and applied directly to prediction cross-sections. The non-perturbative corrections are on the order of 7% for jets of p T < 50 GeV and reduce to zero at higher p T values.
Finally, the inclusive W and Z boson production crosssections in proton-proton collisions are also known to NNLO precision in α s and can be calculated with the FEWZ generator [100]. This generator allows also the prediction of several observables of the final-state objects, such as the rapidity distribution of the produced vector bosons.

Calculations based on resummation
There are specific programs available, such as RESBOS [101], which are based on resummed calculations and therefore are suited to describe the transverse momentum spectrum of vector boson production. RESBOS provides a fully differential cross-section versus the rapidity, the invariant mass and the transverse momentum of the vector boson as intermediate state of a proton-proton collision. The resummation is performed to NNLL approximation and matched to NNLO perturbative QCD calculations at large boson momenta.

Overview and predicted inclusive cross-sections
A summary of all Monte Carlo (MC) generators used to describe the relevant signal processes in this work is shown in Table 1. The order of perturbation theory, the parton shower matching algorithms and the corresponding physics processes are also stated. Table 2 summarises several predictions for different generators and PDF sets for production cross-sections of selected final states in proton-proton collisions at √ s = 7 TeV. Uncertainties due to scale and PDF variations are shown in addition. As indicated in the table, the increase of the crosssection from leading-order to NNLO predictions is more than 15%. The difference between different PDF sets is in the order of 1.5% and covered by the associated PDF uncertainties.

QCD dynamics and angular coefficients
To leading order, the angular distribution of the decay products in the process e + e − → µ + µ− can easily be calculated and exhibits a (1 + cos 2 θ ) dependence, where θ is the angle between the incoming electron and the outgoing positive charged muon. A similar angular dependence is derived for the quark/antiquark annihilation including a Z boson exchange in the corresponding s-channel diagram. However, the coupling and gauge structure of the weak interaction as well as higher-order corrections leads to new angulardependent terms in the differential production cross-sections. The measurement of these terms therefore provides not only Table 1 Monte Carlo programs which are used for the analyses described in this review article. The information on the order of O(α s ) in the matrix-element calculation, the generator functionality, the possibility to match matrix-element calculations with parton showers and the functionality within the analyses are given. The structure of the table is based on [37]. integral rates and shapes Table 2 Prediction of the cross-sections of W and Z boson (66 GeV < m ll < 116 GeV) production in √ s = 7 TeV pp-collisions at leading order, NLO and NNLO in α s calculated by the FEWZ generator. The given uncertainty is calculated for the NNLO cross-section and includes PDF-and scale-uncertainties. an important test of perturbative QCD but also of the fundamental properties of the electroweak sector, as described in more detail in the following paragraphs. The discussion starts with two definitions of rest frames that allow for the definition of the angle θ in proton-proton collisions (Section 2.8.1). Then the general form of the differential Drell-Yan cross-section is introduced in Section 2.8.2, while the interpretation of the corresponding angular coefficients is discussed in Section 2.8.3.

Collins-Soper and helicity frame
The direction of the incoming particles and anti-particles is known in an e + e − collider and hence the reference axis for the definition of the angle θ can be defined in a straightforward way. The situation is very different at a proton-proton collider such as the LHC. Since the vector boson originates from a qq annihilation, a natural choice would be the direction the incoming quark but this can not be done for two reasons. First, the direction of the incoming quark is not know at the LHC. Second, the incoming partons are subject to initial-state radiation, which leads to a non-negligible transverse momenta, p T , of the vector boson when annihilating that can not be determined for the two interacting partons.
To overcome these problems, the rest frame of the vector boson is typically chosen as rest frame in which the angular distributions of the decay leptons is measured. However, the definition of the axes is this rest frame is still ambiguous. To minimise the effect due to the lack of information about the kinematics of the incoming partons, the polar axis can be defined in the rest frame of the vector boson, such that it is bisecting the angle between the momentum of the incoming protons. The y-axis can then be defined as the normal vector to the plane spanned by the two incoming protons and the xaxis is chosen such that a right-handed Cartesian coordinate system is defined (Figure 12). The resulting reference frame is called Collins-Soper (CS) frame [102].
When measuring θ in an analogous way as in an e + e − collision, the direction of the incoming quark and antiquark must also be known but cannot be inferred on an eventby-event basis. However, this can be addressed on a statistical basis. Vector bosons with a longitudinal momentum p z (V ) have been produced by partons which have significantly different Bjorken-x values. Figure 10 illustrates that large x-values enhance the probability for having valence quarks in the interaction and therefore the corresponding antiquark can be associated to the smaller x-values. Hence the measurement of p z (V ) allows us to assign the longitudinal quark and antiquark directions on a statistical basis. It should be noted that large p z (Z) values also imply large rapidities and therefore the statistical precision for the correct quark/antiquark assignment is enhanced for Z bosons in the forward region.
In summary, the angle θ can be expressed in the CS frames as , where E and p Z are the energy and longitudinal momenta of the first and second lepton. The first term of this equation defines the sign and hence the direction of the incoming quark. As previously discussed, large rapidities enhance the probability for a correction assignment of the direction. The second term of the equation corrects for the measured boost due to the hadronic recoil of the event and defines an average angle between the decay leptons and the quarks.
While angular measurements of the Z boson decays are usually done in the CS frame, measurements of the W boson are traditionally performed in the so-called helicity frame. The helicity frame is also chosen to be the rest frame of the vector boson. The z-axis is defined along the W laboratory direction of flight and the x-axis is defined orthogonal in the event plane, defined by the two protons, in the hemisphere opposite to the recoil system. The y-axis is then chosen to form a right-handed Cartesian coordinate system as shown in Figure 12.

Differential cross-section of the Drell-Yan process
The general form of the differential cross-section of the Drell-Yan process pp → Z(W )+X → l + l − (lν)+X can be decomposed as [103], [104] where θ and φ are the polar and azimutal angles of the charged lepton in the final state 2 in the CS frame to the direction of the incoming quark/antiquark. This decomposition is valid in the limit of massless leptons in a 2-body phase space and helicity conservation in the decay. While the functional dependence of Equation 26 on θ and φ is independent on the reference frame chosen, the parameters A i are frame dependent. When no cuts on the finalstate kinematics are applied, the parameters A i can be transformed from one reference frame to another. Due to the limited detector coverage and additional analysis requirements on the kinematics, the coefficients exhibit an experimentdependent kinematic behaviour. Hence the optimal choice of the reference frame will differ for each analysis.
The angular coefficients A i are functions of the vector boson kinematics, i.e. its transverse momentum, p T (V ), its rapidity, Y V , and m V and contain information about the underlying QCD dynamics. They are subject to modifications from higher-order perturbative and non-perturbative corrections, structure functions and renormalisation and factorisation scale uncertainties. Since the PDFs of the proton impact the vector boson kinematics, the coefficients A i also depend indirectly on the PDFs themselves. The 1-dimensional angular distributions can be obtained by integrating either over cos θ or φ , leading to  Fig. 12 Center-of-mass frame (left), helicity frame (middle) and Collins-Soper frame (right) which can be used to extracted several coefficients independently in the case of small data samples.

Interpretation of coefficients
The (1 + cos 2 θ ) term in Equation 26 comes from the pure leading-order calculation of the vector boson production and decay. The terms corresponding to the coefficients A 0 , A 1 , A 2 are parity conserving, while the terms A 3 to A 7 are parity violating. The A 0 to A 4 coefficients receive contributions from the QCD theory at leading and all higher orders, while the parameters A 5 , A 6 and A 7 appear only in NLO QCD calculations and are typically small. Several studies have been published which discuss and predict these coefficients for hadron colliders [105].
It should be noted that all terms except of A 4 are symmetric in cos θ . In the case of the Z/γ * exchange, A 4 appears also in leading-order calculations as it is directly connected to the forward backward asymmetry A f b via which will be discussed in more detail in Section 6.5. The A 4 parameter also plays a special role in the case of W boson polarisation measurements. As discussed in the previous section, vector bosons tend to be boosted in the direction of the initial quark. In the massless quark approximation, the quark must be left-handed in the case of the W boson production and as a result W bosons with large rapidities are expected to be purely left-handed. For more centrally produced W bosons, there is an increasing probability that the antiquark carries a larger momentum fraction and hence the helicity state of the W bosons becomes a mixture of left-and right-handed states. The respective proportionals are labelled with f L and f R . For W bosons with a larger transverse momenta, the production via a gluon in the initial or final state becomes relevant, e.g. via ud → W + g. Hence the vector nature of the gluon has to be taken into account in the prediction of the production mechanisms. For high transverse momenta, also polarisations in the longitudinal state of the W bosons can appear. This fraction is denoted by f 0 and is directly connected to the massive character of the gauge bosons. The helicity fractions f L , f R and f 0 can be directly connected to the coefficients A 0 and A 4 via where the upper (lower) signs correspond to W + (W − ). In particular, the difference of f L and f R depends only on A 4 as The coefficients A 0 and A 2 also play a particular role in the angular decay distributions, as they are related via the Lam-Tung relation [106]. This relation states that A 0 (p T ) and A 2 (p T ) are identical for all p T if the spin of the gluon is one. In case of a scalar gluon, this relation would be broken. It should be noted that the test of this relation is therefore not a test of QCD theory, but a consequence of the rotational invariance of decay angles and the properties of the quark-coupling to Z/γ * and the W boson. At the Z-pole, the leading-order predictions of the p T dependence of A 0/2 for a gluon of spin one are given by [107][108][109] for the process qq → Z g and by for the process qg → Z q. NLO order corrections do not impact A 0 significantly, while A 2 receives contributions up to 20%.

The LHC and the data collected at
From March 2010 to October 2011, the Large Hadron Collider [110] delivered proton-proton collisions at a center-ofmass energy of √ s = 7 TeV to its four main experiments Table 3 Parameters of the LHC at the end of 2010 and the end of 2011 including an estimate of the average number of interactions per bunchcrossing [115].
2 · 10 32 3.6 · 10 33 Average pp-interactions per bunch-crossing ≈ 1.2 ≈ 10 − 15 Table 4 Overview of recorded integrated luminosity in 2010 and 2011 by the ATLAS and CMS experiments. Also shown is the integrated luminosity which is used for physics analyses.
, CMS [112], LHCb [113] and ALICE [114]. The primary LHC machine parameters at the end of the data taking in 2010 and 2011 are given in Table 3. From 2010 to 2011, the number of circulating proton bunches was increased by a factor of 3.8, the spacing between two bunches was decreased from 150 ns to 50 ns and the beam-focus parameter β * was reduced by a factor of 3.5. This resulted in a significant increase of instantaneous luminosity from L = 2 × 10 32 cm −2 s −1 in 2010 to L = 3.7 × 10 33 cm −2 s −1 in 2011 [115]. The total integrated luminosity delivered to the experiments was L ≈ 44 pb −1 in 2010 and L ≈ 6.1 fb −1 in 2011. The data taking efficiency of ATLAS and CMS, when the detector and data-acquisition systems were fully operational, was above 90% for both years. The recorded integrated luminosity, which was used as the data samples for the published physics analyses for ATLAS and CMS in 2010 and 2011, is shown in Table 4 together with their respective relative uncertainties.
The precise knowledge of the recorded integrated luminosity is a crucial aspect for all cross-section measurements. The Van der Meer methods [116], [117] was applied in total three times in 2010 and 2011 to determine the luminosity for ATLAS and CMS, leading to relative uncertainties below 2%. It should be noted that the luminosity determination is highly correlated between ATLAS and CMS, leading to correlated uncertainties in the corresponding cross-section measurements.
The change in the machine settings from 2010 to 2011 leads to an increase of pile-up noise, which is the occurrence of several independent, inelastic proton-proton collisions during one or more subsequent proton-proton bunch crossings. These additional collisions can lead to a significant performance degradation of some observables which are used in physics analysis. The in-time pile-up, i.e. the additional collisions occurring within the same bunch crossing, can be described by the number of reconstructed collision vertices N vtx in one event. The out-of-time pile-up is due to additional collisions from previous bunch crossings, that can still affect the response of the detector, in particular calorimeters, whose response time is larger than two subsequent bunch crossings. The number of interactions per crossing is denoted as µ and can be used to quantify the overall pile-up conditions. On average, there is roughly a linear relationship between µ and N vtx , i.e. < N vtx >≈ 0.6 < µ >. In 2010, the average number of interactions per collision was µ = 2. The first L dt ≈ 1fb −1 in 2011 had < µ >≈ 6, while < µ > of greater than 15 was reached by the end of 2011. This affects several systematic uncertainties related to precision measurements at the LHC.

Coordinate system
The coordinate system of the ATLAS and CMS detectors are orientated such that the z-axis is in the beam direction, the x-axis points to the center of the LHC ring and the yaxis points vertically upwards ( Figure 13). The radial coordinate in the x-y plane is denoted by r, the azimuthal angle φ is measured from the x-axis. The pseudorapidity η for particles coming from the primary vertex is defined as η = −log θ 2 , where θ is the polar angle of the particle direction measured from the positive z-axis. The transverse momentum p T is defined as the transverse momentum component to the beam direction, i.e. in the x-y-plane. The transverse energy is defined as E T = E sin θ .

The ATLAS detector
The "A Toroidal LHC ApparatuS" (ATLAS) detector is one of the two general purpose detector at the LHC. It has a symmetric cylinder shape with nearly 4π coverage ( Figure 14).
ATLAS has a length of 45m, a diameter of 25m and weighs about 7,000 tons. It can be grouped into three regions: the barrel region in the center of the detector and two end-cap regions which provide coverage in the forward and backward direction with respect to the beam-pipe. ATLAS consists of one tracking, two calorimeter and one muon system, which are briefly described below. A detailed review can be found in [111]. The tracking detector is the closest to the LHC beam pipe and extends from an inner radius of 5cm to an outer radius of 1.2m. It measures tracking information of charged particles in a 2 T axial magnetic field provided by a superconducting solenoid magnet system. In addition, the tracking detector provides vertex information, which can be used to identify the interaction point of proton-proton collision and the decay of short-lived particles. Three technologies are used. The innermost part of the tracking detector consists of three silicon pixel detector layers. Each pixel has a size of 50 × 400µm, leading in total to 80 million readout channels. The pixel detector provides tracking information up to a pseudorapidity of |η| = 2.5. The same region is also covered by the semi-conductor tracker, which surrounds the pixel detector. It consists of narrow silicon strips in the size of 80µm × 12 cm, which are ordered in four double layers. The outermost part of the tracking detector is the transition radiation tracker which uses straw detectors and covers an area up to |η| = 2.0. It provides up to 36 additional measurement points of charged particles with a spatial resolution of 200 µm. In addition, the produced transition radiation can be used for electron identification.
In the electromagnetic (EM) calorimeter of ATLAS, the energy of incoming electrons and photons is fully deposited to the detector materials and can be precisely determined. Moreover, the ATLAS calorimeter can measure the location of the deposited energy to a precision of 0.025rad. Liquid argon is used as active material, while lead plates act as absorbers. The absorbers are arranged in an accordion shape which ensures a fast and uniform response of the calorimeter. The barrel region covers a range up to |η| < 1.475, the two endcaps provide coverage for 1.375 < |η| < 3.2. A presampler detector is installed in the region up to |η| < 1.8, which is used to correct the energy loss of electrons and photons in front of the calorimeter.
The hadronic calorimeter ranges from r = 2.28 m to r = 4.23 m and measures the full energy deposition of all remaining hadrons. The barrel part is the so-called tile calorimeter and covers a region up to |η| < 1.0. An extended barrel detector is used for the region 0.8 < |η| < 1.7. Scintillating plastic tiles are used as active medium. Iron plates act not only as absorber material, but also as the return yoke for the solenoid magnetic field of the tracking detector. The granularity of ∆ φ × ∆ η = 0.1 × 0.1 determines the position information of the measured energy deposits, which is roughly 0.1 rad. The hadronic endcap calorimeter covers a pseudorapidity range from 1.5 < |η| < 3.5, where liquid argon is used as the ionisation material and copper as the absorber. The very forward region from 3.1 < |η| < 4.5 is covered by the forward calorimeters, which also uses liquid argon with copper and tungsten as absorbers. Electrons and photons are also detected in the forward calorimeters, as no dedicated electromagnetic calorimeter is present in that region.
The muon spectrometer is not only the largest part of the ATLAS Experiment, ranging from r = 4.25m to r = 11.0m, but also its namesake. Three air-core toroidal magnets provide a toroidal magnetic field with an average field strength of 0.5 T. Muons with an energy above ≈ 6 GeV that enter the toroidal magnetic field will be deflected. This deflection is measured in three layers of tracking chambers. In the barrelregion (|η| < 1.0) and partly in the endcaps up to |η| < 2.0. Monitored drift-tube chambers provide the precise tracking information of incoming muons. For 2.0 < |η| < 2.7, cathode strip chambers with a higher granularity are used. The trigger-system of the muon spectrometer is based on resistive plate chambers in the barrel region and by thin gap chambers in the endcap. Since the ATLAS muon system is filled with air, effects from multiple scattering are minimised. In addition, the long bending path of the muons provides an accurate measurement of their momentum.
The trigger system of the ATLAS detector has three levels. The first level is a hardware based trigger, which uses a reduced granularity information of the calorimeters and the muon system. It defines so-called regions-of-interest, in which possible interesting objects have been detected, and reduces the event rate to ≈ 75 kHz. The second level trigger is software based and has the full granularity information within the region-of-interest and the inner detector information. By this system, the rates are reduced to 1 kHz. The last trigger level has access to the full event information with full granularity and uses reconstruction algorithms that are the same or similar to those used in the offline reconstruction. The final output rate is ≈ 400 Hz.

The CMS detector
The Compact Muon Solenoid (CMS) detector is the second general purpose detector at the LHC with a similar design as the ATLAS detector. It offers also a nearly full 4π coverage which is achieved via one barrel and two endcap sections ( Figure 15). CMS is 25 m long, has a diameter of 15 m and weights 12500 tons. Most of its weight is due to its namegiving solenoid magnet, which provides a 3.8 T magnetic field. The magnet is 12.5 m long with a diameter of 6 m and consists of four layers of superconducting niobium-titanium at an operating temperature of 4.6 K. The CMS tracking system as well as the calorimeters are within the solenoid, while the muon system is placed within the iron return yoke. We briefly discuss the four main detector systems of CMS below; a detailed description can be found in [112].
The inner tracking system of CMS is used for the reconstruction of charged particle tracks and is fully based on silicon semi-conductor technology. The detector layout is arranged in 13 layers in the central region and 14 layers in each endcap. The first three layers up to a radius of 11cm consist of pixel-detectors with a size of 100µm × 150µm. The remaining layers up to a radius of 1.1 m consist of silicon strips with dimensions 100µm × 10cm and 100µm × 25cm. In total, the CMS inner detector consists of 66 million readout-channels of pixels and 96 million readout channels of strips, covering an η-range of up to 2.5.
The CMS electromagnetic calorimeter is constructed from crystals of lead tungstate (PbWO 4 ). The crystalline form together with oxygen components in PbWO 4 provide a highly transparent material which acts as a scintillator. The readout of the crystals in achieved by silicon avalanche photodiodes. The barrel part of the EM calorimeter extends to r = 1.29 m and consists of 61200 crystals (360 in φ and 170 in η), covering a range of |η| < 1.479. The EM calorimeter endcaps are placed at z = ±3.154 m and cover an η-range up to 3.0 with 7324 crystals on each side. A pre-shower detector is installed in order to discriminate between pions and photons.
The hadronic calorimeter of the CMS detector is a sampling calorimeter which consists of layers of brass or steel as passive material, interleaved with tiles of plastic scintillator. It is split in four parts. The barrel part (|η|<1.3) consists of 14 brass absorbers and two stainless steel absorbers as the innermost and outermost plates. The granularity is 0.087 × 0.087 in the η, φ -plane. Due to the space limitations from the solenoid, an outer calorimeter has been installed. It consists of two scintillators at r = 3.82 m and r = 4.07 m with 12.5 cm steel in between. The endcap calorimeters cover 1.3 < |η| < 3.0 and are made of 18 layers of 77 mm brass plates interleaved by 9 mm scintillators. The η-region from 3.0 < |η| < 5.0 is covered by forward calorimeters, positioned at z = ±11 m. They will also register the energy deposits of electrons and photons in this rapidity range.
The barrel and endcap parts of the CMS muon system consist of four layers of precision tracking chambers. The barrel part covers a range up to |η| = 1.3 and drift-tube chambers are used for the tracking. The tracking information in both endcaps (0.9 < |η| < 2.4) is provided by cathode strip chambers. The muon triggers are based on resistive place chambers, similar to the ATLAS experiment [118]. The CMS trigger system has two levels. The first level trigger is hardware based and uses coarsely segmented information from the calorimeters and the muon system. It reduces the rate to 100 kHz. The second level trigger, called the high-level trigger, is a software-based trigger system which is based on fast reconstruction algorithms. It reduces the final rate for data-recording down to 400 Hz.

Reconstructed objects
Measurements of single vector boson production using AT-LAS and CMS data involve in general five primary physics objects. These objects are electrons, photons, muons, neutrinos, whose energy can only be inferred, and particle jets, which originate from hadronised quarks and gluons. An overview of the ATLAS and CMS detector performance for several physics objects is summarised in Table 5.

Electron, photon and muon reconstruction
Electrons candidates are identified by requiring that significant energy deposits in the EM calorimeter, which are grouped to so-called electromagnetic clusters, exist and that there is an associated track in the tracking detector. The transverse momenta of the electrons are calculated from the energy measurement in the EM calorimeter and the track-direction information is taken from the tracking detector. A series of quality cuts are defined to select electron candidates. These cuts include cuts on the shower-shape distributions in the calorimeter, track-quality requirements and the track-matching quality to the clusters. Stringent cuts on these quantities ensure a good rejection of non-electron objects, such as particle jets faking electron signatures in the detector. ATLAS has three different quality definitions for electrons, named loose, medium and tight [119] and CMS analyses use two definitions, called loose and tight [120].
For some analyses in both ATLAS and CMS, electron clusters in the transition region between the barrel and endcap sections are rejected, as cables and services in this detector region lead to a lower quality of reconstructed clusters. These regions are defined as 1.37 < |η| < 1.52 and 1.44 < |η| < 1.57 in ATLAS and CMS respectively. Electron candidates in the forward region from 2.5 < |η| < 4.9 (used by some ATLAS analyses) have no associated track information and therefore their identification is based solely on the shower-shape information.
Photons candidates are reconstructed by clustered energy deposits on the EM calorimeter in a range of |η| < 2.37 and |η| < 2.5 for ATLAS [119] and CMS [120], respectively, as well as specific shower shape cuts. If no reconstructed track in the tracking detector can be associate to the electromagnetic cluster, then the photon candidate is marked as an unconverted photon candidate. If the EM cluster can be associated to two tracks, which are consistent with a reconstructed conversion vertex, the candidate is defined as a converted photon candidate.
Muon candidates are identified by one reconstructed track in the muon spectrometer. Combined muons are required to have in addition an associated track in the tracking detector. The measured 4-momenta, in particular the transverse momenta, of combined muons are based on a statistical combination of the independent measurements within the tracking and muon detectors or a complete refit of all available parameters. For the measurements presented here, the momentum resolution for muons is dominated by the information from the tracking detector for both experiments. CMS can reconstruct muons within |η| < 2.4 [121], while the ATLAS muon spectrometer reaches |η| < 2.7 [122]. However, muons that are reconstructed beyond |η| > 2.5, have no associated information from the tracking detector available and therefore only kinematic information from the muon spectrometer can be used. ATLAS analyses therefore often restrict the muon range to |η| < 2.4.
In many single vector boson measurements, the leptons are required to be isolated meaning that there is no significant energy deposited around the lepton itself. Requiring isolation greatly reduces the number of particle jets which are misreconstructed as electron, photons or muons.
Isolation can be defined in several ways. First, a trackingbased isolation can be used that is defined as where i indicates the sum over all reconstructed tracks in the tracking detector with an energy above a given threshold and within a cone-radius of 0.3 in the (η, φ )-plane. The track from the lepton candidate itself is not considered. This quantity can be normalised by the transverse momentum of the lepton candidate, which ensures a more stable isolation definition for larger transverse momenta. A similar definition can be made using the EM calorimeter, i.e.
where i runs over all EM clusters within ∆ R < 0.3 that are not associated to the reconstructed lepton. ATLAS uses both, tracking-and calorimeter-based isolation criteria as defined above. CMS uses similar isolation variables, but in addition applies non-normalised isolation definitions based on tracks or energies in the hadronic calorimeter.

Hadronic jets and missing energy reconstruction
Hadronised partons are detected as particle jets in the EM and hadronic calorimeters. To reconstruct particle jets in AT-LAS, the energy deposits are merged to topological clusters in a pseudorapidity range of |η| < 4.9. Clusters are seeded by calorimeter cell deposits with a four sigma deviation from the noise level. An anti-k T algorithm [123] is then used to reconstruct the clusters into jets. The typical ATLAS distance parameter of the jet definition is R = 0.4. In addition it is often required that the distance between leptons and jets in the (η, φ )-plane of the detector satisfies the condition ∆ R(l, jet) > 0.3 to avoid double counting [124]. The jet energy and direction is corrected for effects like additional dead material in the detector, the difference in energy response in the calorimeter to hadronic versus EM interactions, the loss of energy outside the jet radius and the presence of energy from pile-up interactions [124]. For CMS analyses, where the detector features a superior tracking system but offers a less precise calorimetry system, the so-called particle flow technique [125] is used. This method combines information from all detector systems, in particular the calorimeter and the tracking detector and aims to identify all particles in mutual exclusive categories: muons, electrons, including the identification of bremsstrahlung photons, converted and unconverted photons, charged and uncharged hadrons. Thus, a full event description of each particle is attempted. This event description is used as input to an anti-k T algorithm with a typical distance parameter of R = 0.5. The jet energy and direction is also corrected for effects like the presence of pile-up and the non-compensation of the calorimeter [126]. Both experiments have dedicated algorithms to identify particle jets originating from band c-quarks. These algorithms combine information about the impact parameter significance of tracks with the expected topology of semileptonic bor c-decays [127].
The energy of neutrinos, which leave the detector unseen, must be inferred though missing energy. While the initial collision energy in beam direction of the partons that are involved in the hard scattering process is not known at hadron colliders, the vector sum of all transverse momenta and energies in the initial state must be zero. Due to energy and momentum conservation, this must also hold for all final state objects in the transverse plane, defined as where the first term corresponds to the vector sum of all transverse energy deposits in the calorimetric system, the second term corresponds to the transverse momenta of the muons reconstructed by the muon systems and the last term corresponds to the transverse momentum sum of all neutrinos in the final state. The latter term is called the missing energy, E miss t,i and its absolute value is called missing transverse energy E miss T . The ATLAS measurement of E miss T uses all electromagnetic and hadronic energy clusters up to |η| = 4.9. Cells which are associated with a particle jet are calibrated with a hadronic energy scale correction, while cells associated to electromagnetic showers are calibrated via the electromagnetic energy scale [124]. The CMS measurement of E miss T follows similar lines, but again uses the information provided by the particle flow algorithms to improve the measurement [125].

Detector simulation and calibration
A detailed simulation of the ATLAS and CMS detector response has been developed over recent years. Both simulations are based on the GEANT4 package [129], which offers the possibility to describe the interaction of all finalstate particles with the detectors at a microscopic level. In a second step, the digitisation of the simulated detector interactions is performed and the nominal data reconstruction algorithms are applied.
Several methods are used to calibrate the detector and to compare data to the simulated events. One important calibration for lepton and jet reconstruction is based on the study of the leptonic decays of the Z boson, which will be briefly summarised in the following. More details on lepton and jet calibration can be found in [126], [124].
The lepton reconstruction and identification efficiencies can be determined in data via the tag-and-probe method. This method makes use of well known decay properties of a resonance, e.g. the Z boson, into two well identified particles. One particle is selected with a strict selection (the tag) to obtain a low background rate. The second particle (the probe) is required to only pass loose selection cuts and can then be used to determine the selection efficiency for tighter requirements.
A simple example is the reconstruction efficiency of muons in the muon spectrometer. The Z boson decays into two muons, resulting in two oppositely charged tracks in the tracking and muon detectors. The tag object is required to have a track in both detector systems. The probe object is required to have only a track in the tracking detector which forms an invariant mass close to the Z boson mass. This ensures a rather clean sample of Z boson events in the muon decay channel. The corresponding reconstruction efficiency can be determined by testing if a matching track to the probe can be found.
Since the Z boson mass and width is precisely known from the LEP experiments, it can also be used to calibrate the energy scale and resolution of leptons. Here, the invariant mass spectra of the leptonic Z boson decays are compared in data and simulations. The peak of the mass distribution is sensitive to the lepton energy/momentum scale, while the width of the distributions gives a handle on the energy and momentum resolutions.
The production of the Z boson also offers the possibility to calibrate the energy scale of particle jets and the hadronic activity in the calorimeters. Z bosons which are produced with large transverse momenta must balance this momenta with additional partons in the final state. The transverse momenta of the Z boson can be reconstructed rather precisely by the 4-momentum measurements of its decay leptons. This transverse momentum must be balanced by reconstructed particle jets, or to be more general, by the total measured hadronic activity. Hence the energy scale and the resolution of particle jets can be calibrated in data.
It should be noted that these methods rely on the available statistics of the corresponding control samples, e.g. on the available number of recorded Z boson events in the leptonic decay channel. While the uncertainties on the detector calibration are usually treated as systematic uncertainties in the physics analyses, they have a significant statistical component which can be reduced by studying more data. Analyses which are based on the 2010 data have therefore significant larger uncertainties due to the limited statistics of the calibration samples. Table 5 Reconstruction performance of electrons, muons and particle jets of the ATLAS and CMS experiment. The reconstruction efficiency and momentum/energy resolutions are shown for the kinematic ranges defined. Further details can be found in the references given.

ATLAS Object definition and algorithm kinematic range reconstruction efficiency p T -resolution Reference
Electron medium quality definition CMS Object definition and algorithm kinematic range reconstruction efficiency p T -resolution Reference 4 Production Cross Section Measurements at the LHC For experimental measurements, the production cross-section is calculated via the following equation The number of signal events is determined as N signal = N data − N bkg , where N data is the number selected events in data and N bkg is the number of background events surviving the signal selection. The factor ε is the efficiency of the signal events passing the signal selection criteria. To correct the crosssection for the choice of a specific decay channel, a branching ratio factor, BR is applied. These ratios are known to a high accuracy for the gauge bosons from LEP experiments [35]. Finally, L dt is the integrated luminosity, which is a measure of the size of the data sample used. The efficiency correction factor ε is usually estimated with simulations of the signal process. These simulations include both a detailed description of the object reconstruction in the detector, called the reconstruction level, and the final-state particle information of the generator calculations, called the generator level. The same signal selection cuts as applied to the data can be applied to the simulated events at reconstruction level. However, the simulation do not model the data perfectly and these differences are corrected in the estimation of ε, following the methods described in Section 3.6. In addition, basic signal selection cuts, such as minimal p T cut, can also be applied to the final-state particles at the generator level. The object value for the final-state particles though differs from the reconstructed quantity. Following these definitions, ε can be defined as the ratio of all events which pass the signal selection on reconstruction level N selected reco.
over the number of all generated events N all gen. . The efficiency correction ε can further be decomposed as the product of a fiducial acceptance, A, and a detector induced-correction factor, C, i.e. ε = A ·C. The fiducial acceptance is the ratio of the number of events that pass the geometrical and kinematic cuts in the analysis on generator level (N selected gen. ) over the total number of generated events in a simulated sample of signal process (N all gen. ). These selection cuts on generator level usually require geometrical and kinematic constraints close to the cuts applied on the reconstructed objects, e.g. leptons in the final state should fulfil p T > 20 GeV and |η| < 2.5. The dominant uncertainties on the fiducial acceptance are the scale and PDF uncertainties.
The detector correction factor C is defined as the number of selected events in simulated sample (N selected reco. ), which now includes a detailed simulation of the detector response, over the number of events in the fiducial phase space at generator level (N selected gen. ). Hence the product of A ·C can be written as N all gen. .
The uncertainties associated with the detector correction factor are dominated by experimental sources, such as limited knowledge of reconstruction or cut efficiencies and the accuracy of the energy/momentum measurements. This factor can be larger than unity due to migration effects from outside the fiducial region into the reconstructed sample. However, in practice this is rarely the case, as detector inefficiencies and the selection criteria on reconstructed objects reduce the number of events.
Defining ε as A ·C is convenient because if the definition of the fiducial volume used for N selected gen.
is close to the cuts applied to the data, this factorisation allows for a separation of theoretical and experimental uncertainties. The fiducial acceptance, A, is completely independent of the detector response whereas the detector correction factor, C, is largely independent of theoretical modelling uncertainties.
In many experimental measurements, the fiducial crosssection, defined as is therefore used, as this definition is only affected to a small extend by theoretical uncertainties. Using fiducial cross-sections has the added benefit that experimental data can be more easily compared to future theory predictions with improved theoretical uncertainties. On the other hand, it should be noted that the theoretical predictions for fiducial cross-sections are also subject to sizeable PDF and scale uncertainties.
In addition to measurements of the inclusive cross-section, measurements of the cross-section as a function of one or more observables can also be made. For a given range of a single observable, this is expressed as where x is the observable being measured, N(∆ x) data , N(∆ x) bkg and C(∆ x) are the same as defined above but for a specific range of x. Differential cross-section measurements allow for a comparison of distributions of the theoretical predictions to the data. The challenge of measuring the differential cross-section is the transformation of the measured distribution, which is distorted by the limited resolution and efficiencies of the detector, to the underlying or true distribution. One possibility to infer the true distribution from the measurement is to directly use Equation 42, which is known as bin-by-bin unfolding. However, this method is only a valid approach if the purity, defined as the ratio of events which fall in the same range of ∆ x in both the reconstructed-and generator-level selections over the total number of events that have been generated in the range ∆ x, of the underlying distribution is high, typically above 90%. When the purity is low, C(∆ x) can have large theoretical uncertainties, since simulations are the only means to estimate this migration between bins. To reduce these uncertainties, advanced unfolding methods have to be used. One widely used approach is Bayesian unfolding [130]. Here, the experimental detector effects are represented in a response matrix, whose elements are the probability of an event in the i-th bin at generator level to be reconstructed in the j-th bin at reconstruction level. The bin size is chosen to be wider than the detector resolution effects, aiming at a purity of > 60%. In the first iteration, the response matrix is derived from simulations. It is then multiplied to the measured spectrum, resulting in a first unfolded spectrum of the data. For the n-th iteration, the response matrix is reweighted to the unfolded spectrum of step n − 1 in order to minimise the bias of the initial prediction. Thus the unfolded spectrum becomes insensitive to the original prior. Other unfolding techniques, such as matrix inversion or single value decomposition [131], are also used.

Event Selection and Background Estimates
The event selection of vector bosons is similar for all studies that are discussed in this article. Hence we introduce a general approach for the signal selection and background estimation for W and Z bosons. The discussion is mainly based on the published inclusive cross-section analyses based on the 2010 data sample [132], [133]. However, important differences in the signal selection for other analyses, such as the production of vector bosons in association with jets, are also highlighted. As the event selection is rather technical matter in nature, this section should be understood as an introduction to the basic concepts. Experienced readers might find the relevant information summarised in Tables 6, 7 and 8.

Signal selection and background estimations of Z/γ * events
The experimental signature of Z bosons in the leptonic decay channel are two oppositely charged, isolated and energetic leptons. These leptons stem from the same vertex and form an invariant mass close to the Z boson mass of m Z = 91.2 GeV. An event display of the typical Z → µ µ event candidate, recorded by the ATLAS detector, is shown in Figure 16. It should be noted that the di-lepton final state contains contributions from both Z boson and virtual photons (γ * ) exchange as well as interference. Therefore, the measured production cross-sections are usually given in terms of a combined Z/γ * exchange.
In most ATLAS analyses, the generic Z/γ * → l + l − selection requires two oppositely charged leptons with an invariant mass between 66 GeV < m ll < 116 GeV. Muons are required to have a reconstructed track in both the tracking and muon detectors within |η| < 2.4 and a minimal transverse momentum of p T > 20 GeV. In addition, the muons are required to pass a relative tracking-based isolation requirement based on a cone radius of ∆ R = 0.2. Electrons are required to fulfil |η| < 1.37 or 1.52 < |η| < 2.4 with a minimum transverse energy of E T > 20 GeV. In addition, the medium identification criteria have to be satisfied. ATLAS also uses forward electrons for some analyses, i.e electrons within 2.4 < |η| < 4.9 which have no associated track in the tracking system. For those electrons tight identification criteria have to be fulfilled and at least one of the two signal electrons must be within |η| < 2.5 and have a corresponding track in the tracking detector.
CMS selects Z → µ µ events by requiring two reconstructed, oppositely charged muons within |η| < 2.1 coming from the same vertex and a transverse momenta requirement of p T > 20 GeV. Both muons have to fulfil a relative tracking-based isolation requirement within a cone radius of ∆ R = 0.3 and must match to the corresponding di-lepton  trigger objects. Later CMS analyses, extend the |η|-requirement to 2.4 and lower the cuts on the muon transverse momenta: the leading and subleading muon have to fulfil p T > 14 GeV and p T > 9 GeV, respectively. For the electron decay channel of Z → ee, at least two reconstructed electrons within |η| < 1.44 or 1.57 < |η| < 2.5 and E T > 20 GeV are required. In later CMS analyses, this was relaxed to require the leading electron to have p T > 20 GeV, while the subleading electron must fulfil p T > 10 GeV. While it is not required that both electrons have oppositely reconstructed charge, they must be matched to the corresponding trigger objects. The electrons must also satisfy a relative trackingbased isolation requirements within a cone radius of ∆ R < 0.3. The Z boson mass range for both channels is defined as 60 GeV < m ll < 120 GeV. The chosen mass range of the CMS experiment is larger compared to the ATLAS definition. This leads to an increase of the available phase space by a factor of 1.015 for CMS. The signal selection cuts are summarised in Table 6.
For measurements of Z/γ * production in association with jets, the jets are required to have p T > 30 GeV. For ATLAS measurements, the jets at high rapidities are used, |y| < 4.4, whereas CMS uses jets within the acceptance of the tracker, |η| < 2.4. Since ATLAS and CMS use jet algorithms with different size parameters (see Section 3.5) the jet energies can not be directly compared. Jets within an ∆ R < 0.3 or ∆ R < 0.5 of an electron or muon are not counted for CMS and ATLAS analyses, respectively.
For Z/γ * → l + l − events, the main background contributions stem from Z → ττ events, di-boson events, tt decays and QCD multi-jet events. Z → ττ events can pass the signal selection when the τ-leptons decay into electrons or muons. Di-boson production such as W Z → l ± νl + l − and top-quark pair production such as (tt → W + bW −b → l + νbl − ν) both have signatures with two energetic and isolated leptons. With the exception of the W Z di-boson process, these processes though do not peak at m Z and are largely removed by the mass cut. The QCD multi-jet events do not necessarily have a lepton in the final state and are discussed in more detail below.
To estimate the backgrounds, CMS often uses a datadriven approach, which exploits the fact that most of the mentioned background processes have an eµ decay channel, while the signal has two same flavour leptons in the final state. By requiring opposite flavour leptons, the background can be directly estimated after correcting for differences in the lepton reconstruction. For most ATLAS analyses, simulations are used for these estimates, since these processes are theoretical well understood in both the absolute background contribution as well as the predictions of the kinematic distributions. In Z/γ * production in association with jets, the background from tt production becomes more significant for larger jet multiplicities. In this case, ATLAS analyses use a data-driven approach similar to the CMS method.
The QCD multi-jet background cannot be predicted precisely and must be estimated with data-driven methods. QCD multi-jet events pass the signal selection cuts in one of two ways: a jet is misreconstructed in the calorimeter and fakes an electron signature or the jet contains a heavy-flavour quark or kaon which decays into an electron or muon. In the first case, jets can fake an electron signal without a real electron in the jet itself whereas in the second case a real lepton is present. The main difference between the lepton signatures for QCD multi-jets versus those from Z/γ * events is the isolation properties and -in the case of electrons -the calorimeter shower-shapes. While the leptons in signal events appear very isolated in the detector, jets contain a significant number of adjacent particles. Similarly W boson production in association with jets can also mimic this signature, where one lepton comes from the leptonic W boson decay, and the Table 6 Summary of the kinematic cuts used by the ATLAS [132] and CMS analysis [133] on leptons and their invariant quantities for the electron and muon decay channel of Z and W Bosons respectively.
To estimate these backgrounds, both experiments use similar data-driven approaches. A control region in data dominated by the QCD multi-jet events is used to define the kinematic distributions of the background. For the muon decay channel, this is achieved by inverting the isolation cut of one of the muons. The control region for the electron channel is obtained by requiring a non-isolated electron which only passes the loose electron identification cuts. The uncertainties of the predicted background distributions can be crosschecked by comparing the spectra to same-sign, isolated dilepton events, which is also expected to be dominated by QCD multi-jet background. The absolute normalisation of the QCD multi-jet background is then achieved by adjusting the sum of the expected signal and other background template to the data as a function of the invariant mass. For Z/γ * production in association with jets, the normalisation of the QCD multi-jet background is determined for each jet multiplicity separately.
Some CMS analyses extract the signal yield together with the lepton trigger and reconstruction efficiencies by using a simultaneous fit to the measured invariant mass spectra in several di-lepton candidate categories, e.g. two combined muons or one combined muon and one inner detector track. The shapes of the signal and background distributions are taken from MC predictions or data-driven approaches as described above.
For measurements of a Z boson in association with bjets, backgrounds from Z events in association with light (u, d, and gluon) jets and c-quark jets dominate. To determine the number of Z+b-jet events, first di-boson, single-top, tt, and W → τν or Z → τν backgrounds are removed. For CMS analyses, the dominant background of tt is normalised to the data at large values of the Z boson mass peak, whereas for the ATLAS analysis, simulations are used to subtract the background. To extract the number of Z +b-jets events from the light-and c-jet events, both ATLAS and CMS use a similar approach; a maximum likelihood fit is preformed using a b-tagging observable. For the CMS analysis, this observable used is the invariant mass of the secondary vertex, estimated from the b-tagging algorithm. For ATLAS measurements, the observable used is one of the outputs of the b-tagging algorithm's neural network. In both cases, these observables are chosen because they give good separation between light-, cand b-jet events. The number of Z+b-jets is then determined from the fit.
The selected data for the inclusive Z boson production measurements based on the 2010 data sample for ATLAS and CMS together with their expected signal and background contributions, as well as the respective uncertainties, are summarised in Table 7. Similar background contaminations and associated uncertainties are seen in the analyses which are based on the 2011 data. The invariant mass distributions and the p T spectra of the decay leptons for the selected data samples and the signal MC predictions are shown for ATLAS and CMS in Figure 18 and 19, respectively. All detector correction effects have been applied. The background contribution is a few percent, making the leptonic Z boson decay channel one of the cleanest signatures at the LHC. Hence it is an ideal channel for precision measurements of the Standard Model as well as for the detector calibration. Overall, excellent agreement between data and the predictions can be seen.

Signal selection and background estimations of W boson events
The leptonic decay of the W bosons (W ± → l ± ν) leads to an isolated and energetic lepton and missing transverse energy. An event display of the typical W → eν event candidate, recorded by the CMS detector, is shown in Figure  17. Since no information on the z-component of the missing energy is available, the mass of the W-boson cannot be reconstructed. However, the invariant mass projection to the transverse plane, defined as  can be reconstructed, since the (x, y) components of the neutrino momentum are inferred from the E miss T . This observable is identical to the W boson mass when the decay happens purely in the x-y-plane.
The ATLAS analyses select W boson events by requiring one reconstructed, isolated lepton, a minimal E miss T of 25 GeV and a minimal transverse mass of m T > 40 GeV. For the muon decay channel, one combined reconstructed muon with p T > 20 GeV within |η| < 2.4 is required. In the electron decay channel, electrons are required to fulfil |η| < 1.37 or 1.52 < |η| < 2.4 with a minimum transverse energy of E T > 20 GeV and medium identification criteria.
The signal selection in the corresponding CMS analysis is significantly different. W boson candidate events are selected by only requiring one reconstructed electron with E T > 25 GeV and |η| < 2.5 or one reconstructed muon with p T > 25 GeV and |η| < 2.1. In later analyses, the η requirement has been relaxed to |η| < 2.4 for muons but the threshold for electrons raised to E T > 35 GeV. Events with a second reconstructed lepton with p T > 15 GeV are vetoed. No additional cuts on E miss T and m T are imposed. For the CMS measurement of W events in association with jets, the selection is slightly modified. The cut on the reconstructed lepton is p T > 20 GeV with no additional leptons above p T > 10 GeV. A cut of m T > 20 GeV is applied.
Both experiments use the signed curvature of the lepton tracks in the inner detector to determine its charge and hence also the charge of the W boson. While the charge misidentification is rare in muon events, a significant fraction of electron charges are mismeasured. Due to substantial material in the tracking detector, a large fraction of electrons radiate photons which in turn may convert to electron-positron pairs close to the original electrons, leading to charge misidentifications during the track reconstruction. In addition to the tracking information, CMS also uses the vertex and cluster position of the calorimeter for the charge identification.
For W production in association with jets, the jet selection is the same as described in Section 5.1.
For both experiments, the major sources of backgrounds for the signal selection are Z/γ * production, the τ-lepton decay channel of the W-boson, di-boson production, QCD multi-jet events and top-pair production.
The Z/γ * process can pass the signal signature when one lepton is not reconstructed, e.g. by being outside of the detector acceptance, thereby creating significant amounts of E miss T . This background is theoretically understood to high precision and therefore the kinematic distributions can be predicted directly from simulations. The normalisation is either taken from a control sample (by requiring two reconstructed leptons), or also from simulations. Similarly the W → τν background, where the τ lepton decays further into electrons or muons and di-boson production, where one or both of the bosons decays to leptons, is also theoretically well understood and modelled to a sufficient precision by simulations.
As discussed in Section 5.1, the QCD multi-jet background must be estimated using data-driven techniques. In ATLAS analyses, the QCD multi-jet control region in the muon channel is defined by reversing the isolation and removing the cut on E miss T . For the electron channel, the control sample is defined by inverting some electron identification criteria and not applying an E miss T requirement. The normalisation of the QCD multi-jet background is determined from data using a fit of the E miss T distribution, the results of which can seen in Figures 20 and 21. For W production in association with jets, the normalisation is determined separately for each jet multiplicity.
For measurements of the inclusive W production, CMS extracts the number of W signal events with a binned, extended maximum likelihood fit to the E miss T distributions. The E miss T distributions for the signal and for the Z/γ * , tt and W → τν backgrounds are based on simulations. The shape of the QCD multi-jet background E miss T template is determined in a control region, defined by inverting a subset of the electron identification criteria or the muon isolation requirement for the W → eν and W → µν-channel, respectively. The fit is performed separately for W + and W − signal events.
Similar to Z/γ * and di-boson production, top-pair production is also theoretically well understood but this background is large for W production in association with jets. For inclusive measurements of W production where the toppair production is a small contribution, simulations are used for the background estimates. For W events with jets, CMS uses a data-driven approach to determine simultaneously the number of both the top-pair events as well as the QCD multijet events. This method exploits two features about tt and QCD multi-jet events. First, since tt events contain a semileptonic decay of the W, these events also peak in m T at the W mass. In contrast, QCD multi-jet events do not peak and have a falling m T spectra. Second, tt events also contain jets from b-quarks which can be selected via b-tagging. To determine the normalisation for tt and QCD multi-jet events, a 2-dimensional fit in m T and the number of b-tagged jets is performed in each jet multiplicity bin. For the ATLAS measurements, the number of tt events is determined using a 1-dimensional fit in the rapidity of the lepton as well as the mass of the W-jet system for each jet multiplicity. The fitted number of tt events is consistent with those from the simulations but has a large statistical uncertainty. For this reason, the ATLAS measurements use tt simulations for the background estimates.
For the measurement of W + b-jets from ATLAS where the tt background is kinematically very similar to the signal events, control regions with four jets are used to constrain the normalisation for the tt events. In addition the normalisa-tion of single-top events is constrained by fitting the invariant mass of the W boson and b-jet system. The extraction of the b-quark events from the light-and c-jet background uses a similar approach as outline in Section 5.1.
Again, the 2010 analyses for ATLAS [132] and for CMS [133] are chosen as an example of the expected background contributions. The resulting number of selected events of ATLAS and CMS together with their expected signal and background contributions and their uncertainties are summarised in Table 8. The corresponding E miss T and m T distributions of selected W-boson events is shown for the ATLAS and CMS experiments in Figure 20 and 21 respectively.

Inclusive and Differential Cross-Section Measurements
The Standard Model predictions of the Drell-Yan processes pp → W ± + X and pp → Z/γ * + X can be tested in a completely new energy regime at the LHC: the study of the Drell-Yan processes provides a unique opportunity to test perturbative QCD predictions and improve the knowledge of the proton's PDFs (Section 6.1, 6.2). The measurement of the transverse momentum distribution of vector bosons can be used to test resummation techniques in addition to higherorder corrections in QCD calculation. The corresponding measurements are presented in Section 6.3. Moreover, the measurement of the forward-backward production asymmetry can constrain the vector and axial-vector couplings of the Z boson to the quarks, where the latest results are presented in Section 6.5 and the measurement of W boson polarisation can test the electroweak properties of the underlying production mechanism, discussed in Section 6.6. Finally, the precise understanding of the Drell-Yan processes is a keyelement for the search of beyond the Standard Model signatures at the LHC. Very similar decay signatures of W and Z bosons are for example predicted by models of large extra dimensions [134], additional U(1) gauge groups [135] or quark-lepton compositeness models [136]. Hence, deviations from the predicted production properties could open the window to new physics.

Inclusive cross-section measurements
The inclusive production cross-sections for the Drell-Yan processes for the W boson and Z/γ * exchange are known to NNLO precision in the strong coupling constant to a precision of ≈ 2% (Section 2.7.7). The dominating uncertainties are to due the limited knowledge of PDFs, while scale uncertainties play only a minor role.
At tree level, Z bosons are produced by the annihilation of quarks and antiquark pairs, i.e. uū, dd and to some extent ss. While the u and d quarks are mainly the valence quarks of the proton, their respective antiquarks are always sea-quarks in proton-proton collisions. The situation is different for the production of W ± bosons, since their production mechanism depends on their charge. The dominant processes for W + and W − are ud → W + and dū → W − , respectively. Since two u-valence quarks are available in the proton, but only one d-valence quark, more W + bosons are expected to be produced. The ratio of the W + and W − production therefore allows for a precise test of QCD predictions, as many theoretical and experimental uncertainties cancels in their ratio ( Table 10).
As one of the first measurements performed at the LHC, the cross-sections times leptonic branching ratios σ W ± ·BR(W → l ± ν) and σ Z · BR(Z → l + l − ) of inclusive W and Z production for electron and muon final states were published by both experiments. These measurements, based on the 2010 data sample with L dt = 35 pb −1 , are not limited by their statistical precision [132], [133], but by the knowledge of the integrated luminosity. Hence, the inclusive results, which are discussed in the following, are based solely on the 2010 data sample.
The measurement strategy is based on Equation 39, which was discussed in detail in Section 4: The number of selected signal events is first corrected for the expected background contribution and then for detector effects via a factor C within a fiducial volume. The division by the integrated luminosity corresponding to the analysed data sample results in the fiducial cross-section. This can be extrapolated in a second step to the full inclusive cross-section via the acceptance factor A. The detector efficiency factors C and the acceptance factors A with their respective uncertainties are shown for both experiments and both decay channels in Table 9. The dominating experimental uncertainties are due to lepton scales and efficiencies.
The combined results for the inclusive cross-sections for W ± and Z/γ * for both experiments are also shown in Table  9. The dominating experimental uncertainties are due to the limited knowledge on the integrated luminosity. In fact, by using the theoretical predicted cross-section, the integrated luminosity of a data sample can be estimated.
The luminosity uncertainty on the cross-section measurement cancel in cross-section ratios, as well as some of the experimental and theoretical uncertainties. Hence, also cross-section ratios have been published for ATLAS and CMS. Special focus will be drawn to the σ (W + )+σ (W − ) σ (Z) ratio, shown in Table 10, where all correlated uncertainties have been taken into account. The NNLO prediction of the ratio is also given.
A simple leading-order calculation for the expected crosssection ratio for (W + + W − )/Z highlights the dependence on the quark-distribution functions. Ignoring heavy quark and the γ * contributions, as well as Cabibbo suppressed parts of the cross-section, leads to Table 8 Data sample and background estimations of the ATLAS and CMS inclusive analyses for the process W ± → l ± ν, based on the 2010 data sample.  : The E miss T and m T distributions for the selected W + → l + ν candidate events. The points with the error bars represent the data. Superimposed are the contributions obtained with the fit for QCD multi-jet background (violet, dark histogram), all other backgrounds (orange, medium histogram), and signal plus background (yellow, light histogram). The orange dashed line is the signal contribution.

Table 10
Results of the production cross-section ratio σ (W + )/σ (W − ) and σ (W ± )/σ (Z) from the ATLAS and CMS analyses, based on the 2010 data sample. The ATLAS measurement of σ (W ± )/σ (Z) = 10.89 was extrapolated to the CMS mass-range definition of the Z-Boson. The expected theoretical value is also shown.

ATLAS CMS Theory (NNLO)
where u v (x) and d v (x) are the up-and down-valence quark distributions and u s (x) and d s (x) the respective sea-quark distributions. When assuming that the light sea and antiquark distributions are the same for a given x and considering that (g V (u) 2 +g A (u) 2 ) ≈ (g V (d) 2 +g A (d) 2 ), this reduces to Table 9 Summary of the cross-section results of the inclusive W and Z analyses of ATLAS and CMS based on the 2010 data sample. The combined cross-sections are given with their respective statistical, systematic, acceptance and luminosity uncertainty, respectively. 6.05 ± 0.02 ± 0.07 ± 0.10 ± 0.21 4.16 ± 0.01 ± 0.06 ± 0.08 ± 0.14 0.937 ± 0.006 ± 0.009 ± 0.016 ± 0.032 i.e. only a small dependence on PDFs is expected if the PDFs have been determined with the assumption of same light sea and antiquark distributions in the proton, i.e.q(x) = q(x) for q = u, d, s. As this symmetry assumption is inherent for the main PDF fits, 3 no large difference in the theoretical predictions based on different PDF sets are observed. The good agreement between the measurements (Table 10) and the predictions are a remarkable confirmation of perturbative QCD calculations, but also support strongly the assumption of a flavour independent light quark sea at high scales, where x is small compared to 0.1, i.e.ū ≈d ≈c ≈d at Q 2 ≈ m 2 Z . The above argument does not hold true for charge-dependent cross-section ratios, such as σ (W + )/σ (Z), σ (W − )/σ (Z) and σ (W + )/σ (W − ). They inhibit a significantly larger dependency on differences in the u-and d-quark distribution functions. However, the largest constraints on PDFs do not come from the inclusive cross-section measurements but from differential measurements, which are also discussed in the following section.
In summary, the inclusive cross-section measurements were one of the first published measurements at the LHC, that confirmed NNLO perturbative QCD predictions in a new energy regime. By now, also inclusive measurements of the W and Z cross-section at a center-of-mass energy of 8 TeV are available [137] which are also in very good agreement with the theoretical predictions. 3 Small deviations are included to account for some light sea quark asymmetry near Bjorken x ≈ 0.1.

Differential W boson and Z/γ * measurements
In addition to the inclusive production cross-section, the large available statistics at the LHC also allows for measurements of differential production cross-sections with high precision. Of special importance here is the measurement of the rapidity distribution y V 4 of the vector boson, as it allows for a direct determination of the momentum fractions x 1/2 of the interaction partons, via where m V is the mass of the vector boson. A center-of-mass energy of √ s = 7 TeV allows therefore to reach x-range from ≈ 0.001 to ≈ 0.1 for the study of W and Z bosons. The boson rapidity distributions are calculated up to NNLO in QCD theory [100] and are dominated by PDF uncertainties. Hence the measurement of the differential production cross-section of gauge bosons versus their rapidity distribution will provide additional constraints on the proton's PDFs. The results of deep inelastic scattering experiments provide constraints on the sea quark and gluon distributions at small and medium x values, while the studies of W and Z production at the Tevatron provided important information on the valence quark distributions. The additional information by measurements at the LHC on the valence quark distribution is therefore expected to be marginal. However, the LHC measurements have a significant impact on the strangequark PDFs, as well as on the ratio of u/d-quark distributions as discussed below. 4 The rapidity of a particle is defined as y = 1 2 E+pz E−pz , where E is the particle's energy and p z is the longitudinal momentum w.r.t to proton direction.  Fig. 22 Rapidity Distribution for the leading-order production of Z (left) and W bosons (right) in 7 TeV pp collisions. The relative contribution of the different production channels is also shown.
The rapidity distribution of the Z/γ * → l + l − process can be directly inferred from data, since both four-momenta of its decay-leptons can be precisely measurement. Hence it will give new information on the uū, dd and ss PDFs. While the information on u and d quark distribution is already largely determined by previous experiments, in particular the uncertainty on the strange quark content can be improved. Figure 22 shows the contribution of the different quark/antiquark annihilation processes for different y Z values. While the uū annihilation is dominating in the central region, the dd annihilation process is expected to have a larger influence for larger rapidity values. A precise measurement in the central rapidity region can also give additional constraints on the ss PDFs. In addition, the study of y Z/γ * for different mass intervals can probe different xregimes, e.g. low-mass Drell-Yan events will probe in general small values of x than high-mass Drell-Yan events. Such studies can be used to improve the knowledge on the ratio of uand d-quark distributions.
ATLAS published a combined differential dσ /d|y Z | crosssection in the fiducial region 5 for the electron and muon decay channel of Z/γ * → l + l − based on L dt ≈ 35pb −1 [132]. Figure 23 shows the results including NNLO theory predictions with various PDF sets. The largest rapidity reached is y Z = 3.5 which is due to the inclusion of forward electrons in this study. In addition, ATLAS published a differential cross-section of the Drell-Yan process in the electron decay channel versus the invariant mass of the dielectron pairs, based on the full 2011 data sample [138]. The comparison of data and NNLO predictions with various PDF sets is shown in Figure 24 The latest CMS publication on Z → l + l − [139] is also based on an integrated luminosity of L dt ≈ 4.5fb −1 and L dt ≈ 4.8fb −1 for the muon and electron channels, respectively. The measurement is performed in a double differential way over the mass range of 20 GeV to 1500 GeV and an absolute di-muon rapidity from 0 < |η| < 2.4. The resulting rapidity distributions for three different mass regions are shown in Figure 25, together with the NNLO prediction for various PDF sets. The differential cross-sections have  been extrapolated to the full phase space and normalised to the Z peak cross-section, which is defined in a mass region of 60 − 120 GeV. Hence many systematic uncertainties cancel in the ratio. The dominating remaining uncertainties are due to the efficiency corrections in the muon channel and energy scale uncertainties in the electron channel.
Especially in the low-mass region, sizeable differences between the different PDF sets can be seen. The uncertainties of the available data are small enough to provide sufficient sensitivity to allow for an improvement over the existing PDF sets. The uncertainty on the u/d ratio can be improved by more than 20%.
The W ± boson rapidity distribution is sensitive to the udand dū-quark distribution. Their respective contribution is shown in Figure 22. However, its rapidity distribution in the leptonic decay channel is not directly accessible as the longitudinal momentum of the neutrino not measured. Therefore only an indirect measurement is possible, which is achieved by the measurement of the pseudorapidity of the charged decay leptons η l which are correlated to y W . The production and decay of W boson is described by the V-A nature of the weak interaction. It is therefore expected that the spin of the W boson is aligned with the direction of the antiquark and the charged decay lepton is preferentially emitted opposite to the boost of the decaying boson. The corresponding experimental quantity is the lepton charge asymmetry where l denotes the lepton and dσ /dη the differential crosssection for charged leptons from the W events. The definition of A(η) has the advantage that several systematic uncertainties cancel in its ratio and can constrain the u/d-quark ratio and the corresponding sea-quark densities. Clearly, also the measurement of separate differential cross-sections dσ (W ± → l ± ν)/dη provides the same information when the correlation between the systematic uncertainties is known. CMS published results on the lepton charge asymmetry in the electron and muon decay channels within a fiducial phase space defined by a p T > 35 GeV requirement for the charged decay leptons. Since the study of W → eν [140] is based on only L dt = 0.84fb −1 and the W → µν analyses [141] uses the full available data sample at √ s = 7 TeV, we discuss here only the latter. The A(η) measurement after all corrections is shown in Figure 26 for a minimal muon requirement of p T > 35 GeV. The dominating systematic uncertainties are due to efficiencies and scale uncertainties, as well as uncertainties on the QCD multi-jet background. Statistical uncertainties are small compared to the systematic uncertainties, which range from 0.2% in the central region to 0.4% in the forward region. The correlations between different η-bins are small. The results are compared to NLO predictions for several PDF sets.
ATLAS has published similar results for the full 2010 data sample in both leptonic decay channels within a fiducial phase space, defined by p lep T > 20 GeV, p ν T > 25 GeV and m T > 40 GeV. In addition to the lepton-charge asymmetry, which is shown in Figure 27, also the individual lepton charge distributions for W + and W − have been derived. Similar to the CMS results, statistical uncertainties are negligible compared to the systematic uncertainties.
While most PDF sets show good agreement with data, the MSTW2008 PDF parametrisation has a poor agreement especially in the region of small rapidities. This is due to a problem in d-valence distribution which was fixed in the MSTW2008CPdeutnlo set, which is also shown. Since the uncertainties of the measured A(η) values are smaller by a factor of 2-3 compared to the predicted uncertainties of the studied PDF sets, an improvement of future PDF sets is expected. Some preliminary results which make use if the currently published LHC data, can be found for example in [142]. Figure 28 and 29 show a comparison of the ATLAS and CMS results for the Z/γ * and A W (η) distributions, respectively. We extrapolate the results to a common fiducial volume, defined by 60 GeV < m ll < 120 GeV for the Z/γ * pro- ATLAS has studied the impact of their data on the proton PDFs using the HERAFitter framework [143]. Here, especially the y Z measurement has a large impact on the constraints of the strange-quark PDFs. Even with the limited data sample of 2010, the hypothesis of a symmetric composition of the light sea-quarks at low x [132] is supported. Specifically, the ratio of the strange sea-quark content to the down sea-quark content at x = 0.023 was found to be 1 +0. 25 −0.28 at Q 2 = 1.9 GeV 2 . This is a remarkable results and was confirmed in an improved analyses of the 2011 data sample [144]. So far it has been assumed in most PDF fitting approaches that s =s =ū 2 =d 2 due to the mass difference of the quarks at the starting scale, i.e. before the QCD evolution starts. At higher values of Q 2 , the gluon splitting processes become dominant and lead to a symmetric distribution of sea-quarks. This new results suggests even an equalū−,d-and s-quark content at low Q 2 values [132]. A visualisation of the impacts on the strange quark distributions is shown in Figure 30. The inclusion of electrons in the forward-region of the ATLAS detector extends the available y Z regime and should therefore improve the information on Measured W charge asymmetry as a function of lepton pseudorapidity |η| is shown on the right. All results are compared to the NNLO theory predictions using various PDF sets.   Fig. 28 Comparison of the ATLAS and CMS: Measurements of the normalised differential cross-section of the Drell-Yan process as a function of the rapidity of the di-lepton system. The ATLAS results have been extrapolated to the CMS fiducial volume, as it is significantly reduced in Y ll . For completeness, also the full normalised differential cross-section of the ATLAS experiment is shown. the valence quark distributions. However, the current experimental uncertainties are too large for a significant effect.
The published results on the lepton charge asymmetry for the W boson production impacts moderately the valence quark distribution functions compared to the existing data that include Tevatron results. However, when only using measurements from HERA and LHC, an improvement on the valance quark distributions of more than 30% for the full xrange can be observed, compared to HERA measurements alone.
Also the LHCb experiment published a differential crosssection of the W ± and Z boson production in the forward rapidity region (2.0 < η < 4.0), based on L dt ≈ 37pb −1 , which is consistent with the measurements of ATLAS and CMS. A detailed discussion can be found in [145].
In summary, the double-differential cross-section measurements of the W and Z bosons lead to important constraints to the PDFs of the proton. Figure 31 shows the improvement of theū ands parton density functions with and without including the current available LHC data based on the NNPDF group [146]. In particular the ATLAS analysis [132] suggests that the strange quark content is comparable to theū andd content even at low scales-With measurements using the full 2011 data set, the statistical and systematic uncertainties are expected to be decreased significantly and a further improvement > 20% of the strange-quark PDFs and the u/d-quark ratio is anticipated. It is not clear how much improvement of a similar study of the full 2012 data set at a center-of-mass energy of √ s = 8 TeV can be expected. Even though the data set is larger by a factor of five, the 2011 re-  Comparison ofū-quark (left) ands-quark (right) PDFs from the NNPDF2.3 NNLO sets with no LHC data and with LHC data used in the fit. Further comparison plots can be found in [146] and [142].
sults are not statistics limited and the increase of center-ofmass energy is only 10%.

Transverse momentum measurements of vector bosons
As discussed in Section 2.7.6, the transverse momentum distributions of vector bosons provides an important test of QCD corrections in the initial state of the production process due to the absence of colour flow between the initial and final state. In particular, predictions based on resummation techniques can be tested, which play an important role for the expected p T spectra between 0 and ≈ M V /2 GeV. Only a mild dependence on the proton PDFs is expected. Besides the test of QCD calculations, the accurate understanding of the vector boson's p T spectra is essential for the measurement of the W boson mass at the LHC, especially when the p T spectrum of the decay leptons of W ± → l ± ν is used as a sensitive variable for m W .
The transverse momentum distributions of electron and muon pairs from Z/γ * events can be measured directly with the reconstructed four-momentum information of the decay leptons. The p T (Z) momentum resolution for p T (Z) < 40 GeV is typically ≈ 3 GeV for ATLAS and CMS. A finite binning with a similar size leads therefore to resolution effects, i.e. bin migration, which make a dedicated unfolding procedure necessary (see Section 4).
CMS has published normalised transverse momentum distributions of the Z/γ * process based on the 2010 data sample [147]. The differential cross-section has been normalised to the cross-section integrated over the acceptance region, defined by p l T > 20 GeV, |η| < 2.1 and 60 GeV < m ll < 120 GeV. Both decay channels have been unfolded using an inverted response matrix and show consistent results, as seen in Figure 32.
ATLAS also published an analysis using the full 2010 data sample [148], also using both leptonic decay channels. The differential cross-sections have been normalised to the fiducial cross-section with 66 GeV < m ll < 116 GeV, p l T > 20 GeV and |η| < 2.4. A Bayesian method has been used for the unfolding of the data, where the resulting distribution is shown in Figure 33 in comparison with different MC generator predictions 6 . The dominating systematic uncertainties are due to uncertainties of the momentum scale and resolution uncertainties as well as from the unfolding procedure.
The low p T (Z) domain of the Z/γ * production can be alternatively probed with the Φ * observable, defined as where ∆ Φ is the azimuthal opening angle between the two decay leptons and as the measure of the scattering angle of the positive and negative leptons with respect to the beam [149]. The Φ * observables is highly correlated to p T (Z)/m ll ; a small p T (Z) leads to a large opening angle and hence a small value of Φ * , while a large transverse momenta lead to small opening angles and therefore larger values of Φ * . A typical value of p T (Z) ≈ 100 GeV leads to Φ * ≈ 1. This variable has the advantage that it is solely constructed using track-based directions which are known to much higher precision than their transverse momenta. ATLAS has published unfolded normalised 7 Φ * distributions for three different regions Z boson rapidity regions (|y Z | < 0.8, 0.8 < |y Z | < 1.6 and |y Z | > 1.6) [150]. The resulting distribution is shown in Figure 34 for both decay channels, together with the theoretical prediction based on the RESBOS generator. Both decay channels lead to consistent results and a clear deviation from the theoretical prediction can be observed; these are consistent with the published results on p T (Z). The systematic uncertainties are smaller than the associated statistical uncertainties for all bins. The statistical precision varies 0.3% for Φ * ≈ 0 to 1.6% for Φ * ≈ 2.5. 6 It should be noted that the differences in predictions of different generators could be due to different scale-parameter settings used. 7 The results have been normalised to the same fiducial regime as the corresponding p T (Z) analysis.  Fig. 34 ATLAS: [150] The ratio of the combined normalised differential cross-section to RESBOS predictions as a function of Φ * . The inner and outer error bars on the data points represent the statistical and total uncertainties, respectively. The uncertainty due to QED FSR is included in the total uncertainties. The measurements are also compared to predictions, which are represented by a dashed line, from FEWZ 2.1. Uncertainties associated to this calculation are represented by a shaded band. The prediction from FEWZ 2.1 is only presented for Φ * > 0.1.

Hadronic Recoil
Transverse Plane W Boson Fig. 35 Illustration of the hadronic recoil in W ± → l ± ν events. While the transverse momenta of the Z boson can be directly interfered by its both decay leptons, the hadronic recoil has to be used in the leptonic decay modes of the W boson.
The measurements of the p T (Z) distribution provides important information for the tuning of MC generators, which can then be indirectly transferred to the prediction of transverse momentum distribution of the W boson, p T (W ). However, an explicit measurement of p T (W ) would allow to directly constrain the p T spectrum of the W boson's decay leptons and therefore estimate uncertainties on an associated m W measurement.
ATLAS also published a measurement of the transverse momentum distribution of the W boson p T (W ) based on the 2010 data sample [151]. The p T (W ) cannot be directly measured from its decay leptons due to the neutrino. However, the p T (W ) must be balanced by the hadronic activity induced by QCD corrections in the initial and final state, i.e. The combined electron and muon measurement of the Z boson transverse momentum distribution (points) and the predictions of several PYTHIA tunes and POWHEG interfaced with PYTHIA using the Z2 tune (histograms). The error bars on the points represent the sum of the statistical and systematic uncertainties on the data. The lower portion of the figure shows the difference between the data and the simulation predictions divided by the uncertainty delta on the data. The green (inner) and yellow (outer) bands are the ±1δ and ±2δ experimental uncertainties.
where p T (had) denotes the hadronic recoil ( Figure 35). Hence p T (had) can be used for the measurement of p T (W ), since it reflects the underlying hadronic activity from the hard QCD interactions. The hadronic recoil has several experimental uncertainties (e.g. pile-up) and also a rather poor resolution compared to the reconstruction of leptons. Hence a data-driven model of the relation between p T (had) and p T (W ) has been used. This model is derived from Z boson events, where the p T (Z) can be directly determined via the decay lepton measurements with a sufficiently good resolution. It is then assumed that the dependence of the hadronic recoil to the transverse momentum of the vector boson is the same in W and Z boson events. The unfolding is performed with a Bayesian approach. The resulting differential cross-section, which has been normalised to the fiducial cross-section measured in the phase space defined via p l T > 20 GeV, |η l | < 2.4, p ν T > 25 GeV and m T > 40 GeV, is shown in Figure 36. It should be noted that the poor resolution of p T (had) implies a significantly larger binning to ensure a stable unfolding procedure. The systematic uncertainties of the data-driven modelling of p T (had) during the unfolding procedure dominate the overall uncertainties up to p T (W ) < 75 GeV. Statistical uncertainties start to dominate for larger p T (W ) values. This measurement has not yet been repeated for the 2011 data sample, as the increase of pile-up further reduces the resolution of p T (had) and hence complicates the unfolding procedure when aiming at a similar binning for the low p T (W ) region. The comparison to the prediction based on RESBOS for both, p T (W ) and p T (Z), (Figure 33, 36), shows a significant disagreement and hence will allow for an improved tuning of the underlying MC generators.
A comparison of the published p T (Z) distributions of ATLAS and CMS is shown in Figure 38. We extrapolated the ATLAS results to the CMS fiducial volume, while the CMS results have been corrected to the ATLAS binning to allow for a direct comparison. A slight tension can be observed for p T (Z) < 10 GeV which is not yet significant.
In summary, a new tune of generators like PYTHIA is needed in order to describe the deviations between measurement and simulations. First measurements of the Z boson transverse momentum distribution at √ s = 8 TeV based on a very reduced data set have also become available [152]. The expected measurements of the full 2012 data set at √ s = 8 TeV could also allow to test electroweak corrections which are predicted to become sizeable at large transverse momenta. In addition, the higher statistics will allow for the measurement of angular coefficients in the Drell-Yan production, as introduced in Section 2.8 (Equation 26), in up to three dimensions. Figure 37 shows the POWHEG and MC@NLO prediction of A 0 vs. the transverse momentum of the Z boson. The observed discrepancy between both generators could be due to the difference in the matching scheme of NLO calculations and the underlying parton shower model (see Section 2.7.3). Therefore a precise measurement   of the angular correlations with the available √ s = 8 TeV data sample is mandatory to test this important aspect in modern QCD calculations.

Determination of Γ W
The measurement of the ratio R of leptonic rates for the inclusive production of W and Z bosons at the LHC, as discussed in the previous section, can be written as The inclusive cross-section ratio is known to NNLO in α s and has a numerical value of σ W /σ Z = 3.34 ± 0.08 [153]. The leptonic branching ratio for the Z boson BR(Z → l + l − ) is known to be (3.366 ± 0.002)% from the LEP and SLC experiments [35]. Hence, the measurement of R allows for an indirect determination of the leptonic W boson branching ratio BR(W → l ± ν) = Γ (W ± →l ± ν) . The CMS analysis [133] used the measured ratio R = σ (W ± → l ± ν)/σ (Z → l + l − ) = 10.54 ± 0.19 leading to an indirect determination of BR(W → l ± ν) = 0.106 ± 0.003 In addition, the partial leptonic decay width of the W boson can be calculated within the Standard Model and is given by where corresponding electroweak corrections δ SM l are small, since they are largely absorbed in G F and m W [154]. Therefore the W boson width can be extracted via BR(W → l ± ν) and results in which is consistent with the current world average Γ W = 2085 ± 42 MeV.
A combination of the ATLAS and CMS results on the cross-section ratio could reduce the overall uncertainty by ≈ 20%, when assuming no correlations between the systematic uncertainties. This would lead to an uncertainty on Γ W of 50 MeV and hence would lead to a significant reduction of the uncertainty on the current world average.

Forward-backward asymmetries of the Z boson
The study of the Drell-Yan production at the LHC does also provide information on the weak mixing angle. This measurement is complementary to the Z -pole analyses of the LEP experiments [155], [156].
The differential cross-section for an ff → f f annihilation process 8 at lowest order for a Z boson exchange is given by 8 with f = f since no t-channel contributions should be allowed where cos θ is the angle between the incoming and outgoing fermions 9 , and a f and v f = a f · (1 − 4|q f |sin 2 θ W ) are the axial and vector-axial couplings to the Z boson and q f is the fractional charge of the fermion. Several things should be noted. First of all, the (1 + cos 2 θ ) dependence would also appear in a pure γ exchange diagram. However, the vector-and axial-vector couplings of the Z boson introduce an additional cos θ dependence. Secondly, the differential cross-section depends only on the weak mixing angle sin 2 θ W , when fixing the electric charges, the weak-hypercharges, m Z , and Γ Z for a given center-of-mass energy. By defining forward and background cross-sections in terms of the angle of the incoming fermions, where the angle θ is defined in the CS frame as introduced in Section 2.8.1, a measure for the asymmetry at cos θ = 0 can be defined as also known as the forward-backward asymmetry parameter A FB . At tree level, A FB is given by The measurement at the Z pole provides the most sensitive measurement as Z exchange contributes roughly 100 times more than the γ exchange. Therefore, only small corrections from the interference and pure γ exchange terms are expected. It should be noted that electroweak corrections to sin θ W can be absorbed by defining an effective weak mixing angle sin θ e f f which is therefore used in the actual measurements.
The advantage of the A FB measurement is that it reduces to a good approximation to a pure counting experiment, e.g. by defining where the number of events in the forward and backward regions are labelled as N cos θ >0 and N cos θ <0 . The measurement of A FB can also be used for the search for new physics. While the A FB at the Z boson mass is used for the determination of sin 2 θ W , large invariant masses are governed by virtual photon and Z interference terms. A direct search for a new resonance in the electroweak section 9 or between incoming and outgoing anti-fermions via the study of the invariant mass spectra of di-lepton events might not show an excess if the new resonance has a large width. However, such a new resonance would also interfere with the Standard Model amplitudes and hence introduce a structure in the measured asymmetries A FB near its mass.
Both, ATLAS and CMS have published unfolded A FB measurements and also a determination of the electroweak mixing angle. In addition to the backgrounds from other processes, the measured A FB is diluted by a wrong assignment of the incoming quark and antiquark. This is accounted for in both analyses using simulations and hence relies on a precise knowledge of the PDFs.
The ATLAS analysis is based on the full 2011 data sample and uses the electron and muon decay channels [157]. Since a high acceptance of Z bosons with large rapidities reduces the dilution of falsely identified quark-directions, AT-LAS also includes forward electrons in their analysis. By requiring one electron with |η| < 2.4 and allowing a second electron within |η| < 4.9, an acceptance for Z boson events up to |y Z | < 3.6 is achieved. The A FB is measured in the same fiducial region as the Z boson inclusive measurements. A Bayesian unfolding technique was used to transform the measured raw cos θ * CS distribution in a given mass region to the cos θ * CS at parton level. The unfolding does not only remove detector effects, but also effects from QED radiative corrections. The latter lead to deformations of the di-lepton invariant mass distribution. In order to account for these corrections, the results are unfolded to Born level, i.e. the state, before any emission of final-state radiation. The unfolding is based on a PYTHIA MC sample, where NLO corrections have been applied. QED final-state radiation was accounted for by interfacing the PHOTOS generator. The p T and y distributions of the Z boson have been reweighted to NLO QCD predictions. NLO electroweak corrections have been estimated with HORACE [158]. The resulting A FB distribution for the electron decay channel including one central and one forward electron, is shown in Figure 39, where the results including forward electrons are denoted as CF. Dominating systematic uncertainties arise from the limited knowledge of electron identification efficiencies in the forward region, NLO QCD effects and PDF uncertainties.
The effective electroweak mixing angle was not determined from the unfolded distributions but measured directly from the raw-data distributions. The measured A FB spectra have been compared to MC predictions which have been produced by varying initial values for sin 2 θ e f f . Each prediction was compared to the measured distribution via a χ 2 test. The minimum of the resulting χ 2 distribution yields then the measured sin 2 θ e f f value. The combination for all channels results in sin 2 θ e f f = 0.2297 ± 0.0004(stat) ± 0.0009(syst.) = 0.2297 ± 0.001 , in agreement with the current world average of sin 2 θ e f f = 0.23153 ± 0.00016. The CMS study of the A FB is also based on the full 2011 data sample [159] and the same fiducial volume as the inclusive measurement in both decay channels. The A FB has been measured in four different rapidity regions (0 to 1, 1 to 1.25, 1.25 to 1.5 and 1.5 to 2.4) and ten mass regions ranging from 40 GeV to 400 GeV, leading to 40 measurements in total. The mass spectra of forward and backward events are unfolded independently for each rapidity region using an inverted response matrix approach. The response matrix is based on a NLO MC predictions from POWHEG and PYTHIA. Also the PYTHIA model for final-state radiation has been used. Similar to ATLAS, the unfolding procedure corrects not only for detector effects but also gives the number of forward and backward events on Born level. The resulting A FB distribution for the combination of both decay channels in the most forward region is shown in Figure 40. Dominating systematic uncertainties are due PDFs in the central rapidity regions and due final state radiation modelling uncertainties in the forward region. Overall, a good agreement with the NLO prediction by POWHEG is observed for all kinematic regions.
The measurement of sin 2 θ W by the CMS experiment is based exclusively on the muon decay channel with an integrated luminosity of 1.1 fb −1 [160]. CMS focuses on the muon decay channel because of the smaller background uncertainties and a better understanding of the corresponding detector performance. The actual methodology for the sin 2 θ W measurement is different from ATLAS. CMS uses an unbinned extended maximum likelihood function which is fitted to data in order to extract the effective weak mixing angle. The likelihood function is evaluated on an event-byevent basis and depends on the number of signal and background events and the expected event the probability density functions for the signal and background processes. These probability densities are parameterised as a function of the di-lepton rapidity, the di-lepton invariant mass, their decay angle cos θ * and the weak mixing angle and rely on leading order predictions of PYTHIA, leading order PDF set and a full detector simulation. The impact of NLO effects has been estimated. The minimisation of the likelihood functions leads to a measured value of sin 2 θ e f f = 0.2287 ± 0.0020(stat) ± 0.0025(syst.) = 0.2287 ± 0.003 , also in agreement with the current world average [155]. A summary of the systematic uncertainties of the ATLAS and    Table 11. The dominating uncertainty for ATLAS comes from PDF uncertainties. Also at CMS, theoretical uncertainties due to PDF, FSR and NLO corrections dominate. The remaining experimental uncertainties could be in principle reduced by future studies. Hence a competitive measurement of sin 2 θ W at the LHC relies on a significant improvement of the proton PDFs.

Polarisation measurement of W bosons
The measurement of the angular distribution of W → µν and W → eν events allows for the determination of the W boson polarisation in proton-proton collisions. The theoretical basis of this measurement was introduced in Section 2.8 and relies on Equation 26. This can be rewritten in terms of the fractions of left-handed, right-handed and longitudinal polarised W bosons, f L , f R and f 0 , which is given by where the superscripts ± relate to the charge of the W boson. By definition, f i > 0 and f L + f R + f 0 = 1 must hold. The parameters f i are not expected to be the same for W + and W − in proton-proton collisions, as the ratio of valence u quarks to sea quarks is higher than for valence d quarks.
Hence different angular distributions are expected. The choice of an appropriate reference frame is not trivial, as the W boson rest frame can be only defined by the full four-momentum information of both decay leptons. While the four-momentum information of the muon is present, only the transverse momentum of the neutrino can be measured. In principle, its longitudinal component can be determined through the W mass constraint. However, the corresponding equations lead to two possible solutions thus an unambiguous choice is not possible. Therefore, the measurement is based on a highly correlated variable cos θ 2D , defined as where p T l * is the transverse momentum of the lepton in the transverse W boson rest frame and p T W is the transverse momentum of the W boson in the laboratory frame. The observ-able cos θ 2D can be interpreted as the 2-dimensional projection of cos θ on the transverse plane. The helicity fractions can then be determined by fitting the measured cos θ 2D distributions with a weighted sum of templates obtained from the simulations. Each template corresponds to one helicity state and is weighted by the corresponding f i value.
The ATLAS measurement of the polarisation of W bosons is based on L dt = 35pb −1 of the 2010 data sample [161]. The selection of signal events and the background determination methods are similar to the inclusive measurement. In addition it is required that the transverse W boson mass is within a range of 50 GeV < m W T < 110 GeV. The upper cut was chosen to reject badly reconstructed jets. In order to enhance the polarisation effects, two regions for events with a large transverse momentum of W boson have been defined as 35 < p W T < 50 GeV and p W T > 50 GeV. The MC prediction of the cos θ 2D templates have been obtained independently from MC@NLO and POWHEG. The actual fit was performed using a binned maximum likelihood fit based on MC@NLO templates. The templates predicted by POWHEG have been used for estimating systematic uncertainties. Since only two of the three parameters f i are independent, it was chosen to measure f 0 and f L − f R . The results for p W T > 50 GeV averaged over both decay channels and both charges are shown in Figure 41 together with the expectations from simulations, which agree with the measurement. The systematic uncertainties are dominated by experimental uncertainties on the measured p W T . The CMS measurement is also based on the 2010 data sample [162] and relies on a binned maximum likelihood for to the cos θ 2D variable. The fitting templates for the different helicity states are based on the MADGRAPH generator. In contrast to the inclusive measurements, CMS requires in this case a cut on the transverse mass of m T > 50 GeV for the electrons and of m T > 30 GeV for the muon channel, in order to reduce the QCD multi-jet background. In addition, also a minimal p W T > 50 GeV is required, similar to the ATLAS analysis. The background from tt is reduced by vetoing events with more than three reconstructed jets with a p T > 30 GeV. The resulting fit values for f 0 and f L − f R are shown in Figure 42, independently for both W boson charges.
Both measurements show a clear difference between the left-and right-handed polarisation parameters in proton-proton collisions and are compatible with the Standard Model expectations.

Vector Boson Production in Association with Jets
The cross-sections of W and Z production for different jet multiplicities are an important measure for NLO and MC predictions but the exploratory power of measurements of W +jets and Z+jets production today lies in the precision tests of differential distributions. The cross-section as a function of the jet p T for example is a sensitive test of the scale used in α s calculations whereas the rapidity distributions of the jets is a sensitive test to different PDF sets. Studying the angular distributions between the jets, such as the rapidity differences between two jets, tests hard parton radiation at large angles. Previous publications from CDF and D0 have measured the differential cross-sections as a function of the jet p T , the boson p T , the angular separation between jets as well as other observables [15-17, 163, 164]. However, the kinematic reach of these measurements compared to that at the LHC is limited. For example, at the Tevatron using an integrated luminosity of L dt ≈ 1.7fb −1 , jets up to a p T of 400 GeV can be measured in Z+jets events, while at the LHC using of integrated luminosity of L dt ≈ 4.6fb −1 , jets up to a p T of 700 GeV can be studied. These large data samples at the LHC, allow us to make precision measurements over large regions of the phase space. In addition, we can measure in detail specific topologies, like Z+jets production where the p Z T is greater than 100 GeV or measure differential cross-sections for rare processes, like Z + bb production.
In addition, the last five years have been an 'NLO revolution'. At the time of the LHC turn-on, NLO calculations up to two associated jets were available, while today an NLO calculation up to six associated jets can be achieved [99]. In addition NLO calculations for W and Z production associated with heavy-flavour jets have also expanded greatly in recent years.

Measurements of W +jets and Z+jets cross-sections
At the LHC, W +jets and Z+jets production is dominated by quark-gluon interactions making these measurements different from measurements of the QCD multi-jet process, which is dominated by gluon-gluon interactions. In MC calculations, associated jets to W and Z production can arise either from the matrix-element calculation itself or from quarks or gluons in the parton showering. Jets from the matrix element tend to have a higher p T compared to those from the parton shower and therefore not including multiple partons in the matrix element will result in an underestimate of the jet multiplicity cross-sections. For an excellent review on jets and their properties at hadron colliders, see [165].
Both ATLAS and CMS have performed measurements of Z production in association with jets. The ATLAS and CMS results are based on an integrated luminosity of L dt ≈ 4.6fb −1 [166] and L dt ≈ 4.9fb −1 [167], respectively. Both results measure the jet multiplicity cross-sections up to seven associated jets. Shown in Figure 43 for the AT-LAS measurement, several theory predictions are compared to the data. The ALPGEN and SHERPA generators both include matrix-element calculations that cover up to five partons, with additional jets coming from the parton showering. MC@NLO generates the Drell-Yan process at NLO and includes the real emission of one additional parton and any additional jets from the parton showering. The BLACKHAT-SHERPA results provide fixed-order calculations at NLO for up to four jets. The data is in excellent agreement with BLACKHAT-SHERPA and SHERPA predictions. ALPGEN tends to underestimate the data above five jets, as expected, since above five partons all additional jets in the predictions originate from the parton shower. While MC@NLO agrees with the data for zero-jet and one-jet events, its parton shower model underestimates the observed jet rate by a factor of two. The CMS results are also in good agreement with the BLACKHAT-SHERPA predictions as well as the MADGRAPH predictions, which includes matrix-element calculations up to four partons and the POWHEGBOX predictions, which is a NLO calculation for one jet. The dominant experimental uncertainty in both of these measurements is the uncertainty on the jet energy scale, while the dominant theory uncertainty is due to the scale uncertainties.
Taking the ratio of jet multiplicity cross-section allows for many of the experimental systematic uncertainties to cancel, thereby improving the precision of the measurement. The ratios, R (n+1)/n , shown in Figure 44 for the ATLAS measurement, exhibit a constant or staircase scaling pattern, as derived in Equation 16. As described in Section 2.2.2, basic quantum field theory would predict a Poisson scaling of R (n+1)/n due to successive gluon radiation from an energetic quark. At higher jet multiplicities, though, a constant value of R (n+1)/n is expected due to the non-abelian nature of QCD final-state radiation, i.e. a final-state gluon can radiate an additional gluon. At low multiplicities this constant value is due to a combination of Poisson scaling and parton density suppression, where the emission of the first parton has a stronger suppression than any additional parton. This Poisson scaling can be recovered if the scale difference between the main process (such as Z+1-jet events) and the p T of the second leading jet is large [168]. When requiring the    Similarly, using a smaller data sample of L dt ≈ 35pb −1 , a CMS analysis [169] tested the Berends-Giele scaling hypothesis which similarly states that the ratio can be described as a constant. Since phase space effects can modify this ansatz slightly, a linear function is used C n = α + β n. The Berends-Giele scaling is confirmed to describe the data for events with up to four jets.
Measurements of W production in association with jets has also been performed by ATLAS [170] using an integrated luminosity of L dt ≈ 35pb −1 and by CMS [171] using an integrated luminosity of L dt ≈ 5.0fb −1 . The CMS results, preformed using only W → µν events, measured the cross-section with up to 6 associated jets (Figure 45). Similar to the Z+jets production, the data is in good agreement with the BLACKHAT-SHERPA, MADGRAPH and POWHEG predictions. The Berends-Giele scaling, measured by both ATLAS and CMS [169] at L dt ≈ 35pb −1 , also describes well the W +jets data up to 4-jets.
Using a common phase space, we summarise in Figure 46 the W +jets and Z+jets results from ATLAS and CMS. For the W +jets results the common phase space is one lepton with a p T > 20 GeV and |η| < 2.5, a neutrino from the W de-cay with p T > 25 and the transverse mass of the W of greater than 40 GeV. For the Z+jets results the phase space is defined as two leptons each with a p T > 20 GeV and |η| < 2.5 and invariant mass of 66 GeV < m ll < 116 GeV. For both W and Z production, the jets are defined using an anti-k T algorithm with a distance parameter of R = 0.5, a p T > 30 GeV and a rapidity less than 2.4. The correction factors applied to both the ATLAS and CMS results are derived from SHERPA and range from 1% to 23%. For all jet multiplicities, the correction factors are smaller than the experimental systematic uncertainties.
In summary measurements of the cross-sections of W and Z production in association with jets are in excellent agreement with the predictions. For measurements of the crosssection ratios of R (n+1)/n the experimental uncertainties are much smaller compared to those from the theory predictions. Future measurements should therefore focus on differential measurements of the cross-section to further test perturbative QCD theory.  Fig. 46 Summary of the W +jets (left) and Z+jets (right) cross-sections as a function of the number of jets. Both the ATLAS [166,170] and CMS [169] results have been corrected to a common phase space as described in the text. The cross-sections for each jet multiplicity have been normalised by the inclusive W or Z cross-section. The CMS results shown here use an integrated luminosity of L dt ≈ 35pb −1 , as the most recent CMS results are not yet published. The ratio between the ATLAS and combined CMS results are shown below. Predictions from SHERPA are also shown.

Differential W +jets and Z+jets measurements
Differential measurements of the properties of the jets in W +jets and Z+jets events probe not only perturbative QCD theory, but are also sensitive to renormalisation scales, PDFs and hard parton radiation at large angles. For these measurements, ATLAS and CMS have two major advantages. First, with large data samples very high jet p T and scalar sum scales can be probed. Second, the detectors can measure jets at large rapidities. Both the high p T and large rapidity jet phase spaces have not been extensively measured in the past.
The differential cross-section of both W or Z events as a function of H T is of particular interest. In the CMS measurements H T is defined as the scale sum of all jets passing the selection criteria. In many fixed-order calculations, H T is often used as the value of the renormalisation and factorisation scales. It is also an observable that is very sensitive to missing higher-order terms in theoretical predictions as well as an observable which is often used in searches for new physics. The effect of missing higher-order terms in the predictions can readily be seen in Figure 47, which shows the differential cross-section as a function of H T for W + ≥ 1jet events. At large values of H T , the NLO BLACKHAT-SHERPA predictions underestimate the data. This is because of the limited order of the BLACKHAT-SHERPA calculations which do not include matrix-element calculations of three or more real emissions. Modifying BLACKHAT-SHERPA to include higher-order NLO terms to the N jet ≥ 1 predictions yields good agreement to the data [170]. The SHERPA predictions shows better agreement to the data, compared to MADGRAPH.  The discrepancy between data and predictions in the H T distribution, which is attributed to missing higher jet multiplicities in fixed-order calculations, can be further investigated by comparing the average jet multiplicity as a function of the H T . Figure 48 shows that for higher values of H T 10 , the average jet multiplicity increases. Therefore at large values of H T , a fixed-order calculation for only N jet = 1, will not model correctly the data and agreement to the data can only be restored when including higher jet multiplicities. This conclusion is especially important for searches for new physics which rely on simulations to predict the number of W and Z background events with large values of H T . Using simulations with an insufficient number of partons in the final state will lead to an underestimate the number of W and Z events at high values of H T .
When the p T of the jet is larger than the mass of the Z or W boson, the NLO to leading-order correction factors become large due to QCD corrections, which are of the order α s ln 2 (p T /m Z ). Also at high values of the jet p T , the electroweak corrections, which are usually small compared to QCD corrections, can also reduce the cross-section by 5- 10 ATLAS measurements define H T as the scalar sum of all jets and leptons in the event. 20% for 100 GeV < p ll T < 500 GeV [172]. The differential cross-section as a function of the leading jet p T for Z events with N jet ≥ 1 measured by the CMS experiment is shown in Figure 49. The experimental systematic uncertainties are smaller than those of the theoretical predictions. The NLO predictions from BLACKHAT-SHERPA are consistent with the data, while POWHEG tends to predict slightly harder jet spectra. MADGRAPH also slightly models incorrectly the shape of the data.
The differential cross-sections as a function of the Z boson and jet rapidities for Z+1 jet events was performed by the CMS collaboration [173]. Since Z+jets production involves a relatively high momentum valence quark and a low momentum gluon or quark, the Z boson and jet are usually produced in the same end of the detector, which implies that the rapidity of the jet and the Z boson in one-jet events is highly correlated. Measuring the rapidity sum, y sum = |y z + y jet |, between the jet and the Z boson is therefore sensitive to the PDFs while the rapidity difference y di f = |y z − y jet | is sensitive to the leading-order parton differential crosssection. These results are shown in Figure 50. SHERPA models the data well, whereas the MADGRAPH and MCFM predictions are less consistent with the data. The differences between the different predictions is most clearly seen in the y di f distribution where SHERPA best models the data.
The large data samples of the LHC also allow for precision measurements at high scales. One topological observable of interest in many searches for new physics is the event thrust, which is defined as where the index i is over all jets and the Z boson, p T,i is the transverse momentum of object i and n τ is the unit vector that maximises the sum. In events where the Z and the jet are back-to-back, the thrust is zero. For events with additional jets that are isotropically distributed, the value of the thrust becomes larger. Traditionally the results are presented as ln τ T , so that back-to-back events have a value approaching infinite and isotropic events have a value of −1. CMS measured the differential cross-section as a function of the thrust in two different phase spaces [174]: an inclusive phase space with p Z T > 0 GeV and a phase space region with p Z T > 150 GeV. Similar to measurements of the jet p T , applying a cut on the p Z T tests perturbative QCD theory in a region of phase space where the QCD corrections can be large. As seen in Figure 51, the predictions are within 10-15% of the data in the inclusive phase space, except for PYTHIA which shows large deviations. In the phase space region with large p Z T , the agreement with PYTHIA improves but both PYTHIA and SHERPA tend to predict more events in back-to-back topologies compared to the data.
A measure of the hadronic activity accompanying W production can be investigated by studying the splitting scales in the k T cluster sequence [175]. These splitting scales are determined by the clustering of objects, either calorimeter energy deposits or particle-level hadrons, according to their distance from each other. The final splitting scale in the clustering sequence, called d 0 , is the hardest scale and corresponds to the p T of the jet. Studying the hardest splitting scales is therefore like studying the k T jet algorithm clustering in reverse. Since this algorithm clusters the soft and collinear branchings first, this clustering sequence is akin to studying the QCD evolution in reverse 11   [pb/GeV] T /dp σ d even though the leading-order accuracy for all of these generators should be the same. This study, which also includes measurements that are sensitive to the hadronisation effects and multiple parton interactions, can be used to help tune these generators in the future.
With the large data sets available from the LHC, both the ATLAS and CMS collaborations have studied extensively differential cross-sections for W and Z production in association with jets. These measurements have highlighted a few features. First, at large values of H T the predictions must include a sufficient number of partons in the matrix-element calculation in order to model the data correctly, even at low jet multiplicities. Second, the experimental precision of mea- surements such as the Z boson differential cross-section as a function of leading jet p T can test not only QCD corrections but for the first time become sensitive to the QED corrections in one-jet events. Future measurements with higher transverse momenta of the jets and the boson will be able to better probe these large QCD corrections at these high values of p T as well as be able to make qualitative statements about the accuracy of the QED corrections.
7.3 Measurements of the ratio of W + to W − in association with jets As discussed in Section 6.2, the W ± rapidity distribution is sensitive to the ud and dū quark distributions. In addition the number of W ± events depends on the number of associated jets because the fraction of u and d quarks contributing to the different jet multiplicity processes changes. CMS measured the charge asymmetry defined as for different numbers of associated jets [169]. Since many experimental systematic uncertainties especially the dominant uncertainties due to the jet energy scale cancel in this ratio, the charge asymmetry is a sensitive test even at large jet multiplicities. Figure 53 shows the charge asymmetry for W → µν events. The MADGRAPH predictions agree well with the data, while PYTHIA fails to model the data even for one-jet events.
In order to place constraints on the PDFs, measurements of the W charge asymmetry as a function of the number of jets and as a function of the p T of the W boson are needed. Such measurements at large values of the boson p T could constrain the PDFs at larger momentum fractions x compared to inclusive measurements [176].

Measurements of the ratio of W +jets to Z+jets
Measurements of the W and Z cross-sections in association with jets are plagued by dominant uncertainties from the jet energy scale. Although both CMS and ATLAS have achieved excellent understanding of the jet energy scale, these uncertainties still dominate especially for jets at high rapidities. However, in other cases, such as the measurement of the jet p T (Section 7.2), the experimental measurement is more precise than the theory predictions. Both the theory and experimental uncertainties can be reduced through a measurement of the cross-section ratio between W +jets and Z+jets processes. For example, when comparing ATLAS measurements of W +jets production to the ratio of W +jets to Z+jets production for events with one associated jet, the jet energy scale uncertainty is roughly a factor of two smaller in the ratio measurement.
CMS measured the ratio of W +jets to Z+jets for up to four associated jets [169]. In both the electron and muon channels, the data were in good agreement with the MAD-GRAPH and PYTHIA predictions. ATLAS measured the ratio for exactly one associated jet but for different thresholds of the jet p T [177]. The combined results from the electron and muon channels are shown in Figures 54 and compared to predictions from PYTHIA, ALPGEN and NLO predictions from MCFM. The ratio, which is not constant as a function of the jet p T threshold, decreases because at large jet momenta the difference in the boson masses is small compared to the effective scale of the interaction. All of the predictions model this trend well.
The measurement of the ratio of W to Z production in association with jets is one of the most precise measurements of perturbative QCD. Future measurements, using the full 2011 data set from the LHC will be able to measure this ratio for higher jet multiplicities and as a function of the p T and rapidity of the jets and the H T . This ratio is sensitive to new physics models, especially if the new particles decay preferentially either to the W vs. the Z final state.

Cross-section measurements of W and Z bosons in association with heavy-flavour quarks
The study of W and Z production in association with heavyflavour quarks is of particular importance today. First, the theoretical predictions are less well known compared to the inclusive W +jets and Z+jets predictions. Second, precision measurements of these processes are critical since they are a dominant background in Higgs measurements of W H production with H → bb decays and new physics searches involving heavy-flavour production.
The production of W + b-jet events has two main diagrams at leading order ( Figure 55): W plus a gluon in the final state where the gluon splits to a bb pair, and b-quark in  the initial state where the W is produced from radiation from a quark. The former can be produced both by the matrixelement calculation and by the parton showering model, whereas the latter diagram can only be modelled by including a bquark in the initial state from the PDF. The predictions and measurements of W + b-jet crosssection have a long history. The first measurements by CDF [21] indicated that the measured cross-section was too large by 2.8 standard deviations compared to the predictions. A measurement by ATLAS, using only an integrated luminosity of L dt ≈ 35pb −1 , also reported a larger cross-section by 1.5 standard deviations, while the D0 measurement was consistent with the predictions [24]. An updated ATLAS measurement [178] of the W + b-jet cross-section using an integrated luminosity of L dt ≈ 4.5fb −1 offers the statistical and systematic precision to definitively close this debate. The measurement presented cross-sections for the exclusive one-jet and two-jet final states with a fiducial phase space requirement of at least one b-jet, defined by the pres-ence of a weakly decaying b-hadron with p T > 5 GeV and within a cone radius of ∆ R = 0.3 of the jet axis. The crosssection results are summarised in Figure 56 and compared to calculations from MCFM, POWHEG and ALPGEN. The MCFM predictions are calculated using the 5-flavour scheme (5FNS) which accounts for the presence of b-quarks in the PDF. The ALPGEN and POWHEG predictions use the fourflavour scheme (4FNS). For one-jet events, the measured cross-section is consistent with 1.5 standard deviations to the NLO MCFM predictions, while for two-jet events the measured cross-section is in good agreement with the predictions. In the one-jet case, the difference between data and the predictions can be more clearly understood in the differential cross-section measurement as a function of the jet p T shown in Figure 56. The MCFM and ALPGEN predict a softer jet p T spectrum with respect to the data.
In a complementary result, the W + bb final state was measured by CMS [179], by requiring events with only two jets, both of which must originate from a b-hadron. The measured cross-section of 0.53±0.05(stat)±0.09(syst)±0.06(th)±0.02(lumi)pb is in excellent agreement with the MCFM prediction of 0.52± 0.03pb Similar to measurements of W +b-jet, the Z +b-jet crosssection is much less studied compared to the inclusive Z measurements. Unlike W + b-jet which only has two main contributing diagrams at leading order, Z + b-jet production includes additional leading-order diagrams of Z radiation from an initial-state b-quark and Z radiation from a final-state bb pair ( Figure 57). Predictions which include diagrams with an initial-state b are therefore necessary. CMS presented a measurement of the Z + b-jet cross-section with exactly one b-jet, with at least one b-jet and with at least two b-jets, using an integrated luminosity of L dt ≈ 5 fb −1 [180]. The results were compared to MADGRAPH, aMC@NLO and MCFM predictions. While the measured cross-sections were found to be in fair agreement with MADGRAPH and aMC@NLO, the MCFM results differ by approximately two standard deviations from the data. ATLAS also measured the Z + bjet cross-section using a smaller data sample with an integrated luminosity of L dt ≈ 35pb −1 [181]. The NLO MCFM predictions as well as ALPGEN and SHERPA were found to be consistent with the data but there were signs of tension especially between the ALPGEN and SHERPA predictions themselves. These results are summarised in Figure 64. Updated results with better statistical precision and predictions with massive quark models are needed here to help resolve these differences.
In both W + b-jet and Z + b-jet production, diagrams with a gluon splitting to a bb pair contribute to the matrixelement calculations and in the parton showering model. In a generator like ALPGEN, this overlap is removed by applying a ∆ R cut, so that at small values of ∆ R the gluon split- ting is handled by the parton shower, while at large values it is predicted from the matrix element. The theoretical uncertainties describing collinear b-quark production are large. To test this transition from parton shower to matrix-element calculations, measurements of the gluon splitting at small values of ∆ R is an important topic at the LHC today. This is especially important for new physics searches and Higgs measurements that select b-jets, since high p T b-jets tend to be produced via gluon splitting.
While measurements of gluon splitting at small values of ∆ R are interesting, it is experimentally challenging to measure since the two b-quarks are often reconstructed within the same jet. CMS presented a new approach for this measurement by measuring Z events with two b-hadrons [182]. As the b-hadrons can be reconstructed from displaced secondary vertices, only tracking information is needed and there is no dependence on a jet algorithm. The angular resolution is ∆ R ≈ 0.02 between the two b hadrons. The differential cross-section measurement as a function of ∆ R of the two b-hadrons is shown in Figure 58. The collinear region (∆ R < 0.5) is best described by ALPGEN, while MAD-GRAPH and aMC@NLO predictions tend to underestimate the data. In addition, the differential cross-section was measured for a phase space region where p Z T > 50 GeV (Figure 58). In this phase space, the relative fraction of events with collinear b-quarks increases. Again, ALPGEN gives the best description of the data.
The cross-section of the W boson in association with cquarks, which have sensitivity to the s-quark contribution in PDFs, have not been experimentally well measured in the past. At the LHC, the dominant W+c production takes place via the reaction sg → W − + c andsg → W + +c, as illustrated in Figure 59. Due to the high production rates at the LHC, measurements of the W + c cross-sections have for the first time, sufficient precision to constrain the s-quark in the PDFs at x ≈ 0.01. In the future, measurements of W + c may also be able to help determine if there is a asymme-  try between the s ands sea 12 as suggested by the NuTeV measurements [69,183,184]. Since the c-quark and the W have opposite charge, the W+c production can be measured by subtracting events with the same-sign charge from events with opposite-signed charge. This subtraction will have no effect on the W − + c process, but all other background such as W − + cc and W − + bb are symmetric in same-sign and opposite-sign events and will be removed. In the CMS analysis [185], the jets originating from a c-quark are selected in one of three ways: a selection of a D ± → K ∓ π ± π ± decay by requiring a displaced secondary vertex with three tracks and an invariant mass which is consistent with the D ± , a selection of a D 0 → K ∓ π ± decay by requiring a displaced secondary vertex with two tracks which is consistent with the D 0 , and semileptonic c-quark decay by requiring a muon matched to a jet. The ATLAS analysis [186] selects W + c events by reconstructing the D ± → K ∓ π ± π ± and D * ± → D 0 π ± decay modes or by identifying jets with a semileptonic c-quark decay.
The ATLAS and CMS results both presented measurements of the W + c cross-sections, the cross-section ratio of W + + c to W − + c as well as the cross-sections as a function of the lepton η. As summarised in Figure 60, the measured W + c cross-section in the ATLAS results is most consistent with PDF sets with a relatively higher s-quark density, while the CMS results are most consistent with PDF sets with a relatively lower density. However, the precision of the measurements is not sufficient to make any definitive conclusions. Overall for both the ATLAS and CMS measurements, there is good agreement between the experimental results and the predictions.
In the ATLAS results, the ratio of the strange-to-down sea-quark distribution, r s = 0.5 (s +s)/d, as a function of x is treated as free parameter in the HERAPDF1.5 PDF fits and all other eigenvectors in the fit are constrained within the uncertainties from the HERA data. As seen in Figure 61, the ATLAS results support the hypothesis of an SU(3)-symmetric light-quark sea and are consistent with the results from the ATLAS-epWZ12 PDF fits [144] where the ATLAS W and Z cross-section measurements are included in addition to the HERA data (see also Section 6.2).
In summary with the large LHC data sets, precision measurements of differential cross-sections of W and Z production in association with heavy-flavour quarks can be made for the first time. Measurements of the W +b-jet cross-sections have indicated that the jet p T spectra is not well modelled by the predictions. In addition the predictions for the Z + b-jet cross-section are in tension both with the data and with each other. Future measurements of the differential cross-sections are needed to resolve this. Finally, both measurements of the W + c-jet production from ATLAS and CMS agree with a wide range of PDF sets. Although the ATLAS and CMS results tend to prefer PDFs with a different s-quark density, additional measurements with more data are needed to study this in greater detail. , compared to four different PDF sets: MSTW2008, CTEQ10, NNPDF23 and NNPDF23coll. The ATLAS results are also compared to HERAPDF1.5 and HERAPDF1.5 but including the ATLAS W and Z data, called ATLAS-EPWZ12. The NNPDF23coll PDF set is like NNPDF23 but excludes all fixed target data. In the ATLAS figure, "OS-SS" refers to the subtraction of events with opposite-signed charges and same-signed charges. The predictions in the ATLAS results are made using the aMC@NLO generator, while the predictions for the CMS results use the MCFM generator. : Ratio of strange-to-down sea-quark distributions as a function of x as obtained from the ATLAS-epWZ12 PDF set, the ratio as assumed in the HERAPDF1.5 PDF set and the ratio obtained from the HERAPDF1.5 PDF set but including the ATLAS W + c measurements. The ATLAS-epWZ12 PDF set includes the AT-LAS W and Z cross-section measurements in addition to the HERA data. The error band on the ratio including the ATLAS measurements represents the total uncertainty. 7.6 Electroweak production of Z bosons Z bosons in associations with jets can be produced not only via the Drell-Yan process, but also via electroweak processes, as illustrated in Figure 62. Electroweak processes are here defined as all processes which lead to a final state of two leptons, two quarks and involve the exchange of electroweak bosons in the t-channel. Of special importance is the vector boson fusion process, shown in the first diagram, as it is an important input for Higgs boson studies and the study of electroweak gauge couplings. In a full calculation of the production cross-section of process, which involves all diagrams, large negative interference exists between the pure vector boson fusion process, the bremsstrahlung and the nonresonant (or multi peripheral) processes. The experimental signature of the electroweak production described above, is the typical Z boson topology of two oppositely charged leptons close to the Z boson mass and in addition two high energetic jets. As the momentum transfer to the interacting initial partons caused by the electroweak bosons in the t-channel is relatively small, these jets tend to be produced in the forward region of the detector. The dominant background is due to Drell-Yan production of the Z boson in association with jets. Significantly smaller background contributions are expected from the top-pair production and di-boson processes such as WW , W Z and ZZ.
The CMS experiment has analysed the full 2011 data set to measure the cross-section of the electroweak production of Z bosons [187]. In addition to a standard Z boson selection in the electron and muon decay channels, further cuts on two reconstructed jets within |η| < 3.6 are imposed in order to reject the Drell-Yan background: The transverse momenta of the two jets are required to be p T > 65 GeV and p T > 40 GeV, respectively. In addition, the Z boson rapidity in the rest frame of the two jets has to fulfil |y * | < 1.2. These requirements lead to a signal efficiency of 0.06 and to a signal over background ratio of S/B ≈ 0.1. This low S/B ratio requires the usage of maximum likelihood fits or multivariate techniques for the cross-section determination. Both approaches have been used in [187]. The distribution of the invariant mass of both jets is shown in Figure 63 for the signal-and-background processes. It is the basis of a likelihood fit based on Poisson statistics where the normalisation of the background distributions and the signal distributions are kept as free parameters. The signal extraction via multivariate techniques uses a boosted decision tree approach, which is based on kinematic variables of the jets and leptons and their combinations. Both approaches lead to consistent results, where the multivariate approach provides lower uncertainties. The final measured cross-section within a fiducial phase space 13 is σ meas. EW,Z = 154 ± 24(stat) ± 46(syst) ± 26(th) ± 3(lumi) fb and is in good agreement with the theoretical expectation of σ EW,Z = 166fb. The dominating uncertainties are due to the modelling of the background distributions and the jet energy scale. A possible electroweak Z production can therefore be observed with a 2.6σ significance already within the 2011 data set. The ATLAS experiment has also published a measurement with improved systematic uncertainties based on the 2012 data set at √ s = 8 TeV [188] which confirms the observation of the electroweak production of Z above a 5σ confidence level. Measurements at a higher center-of-mass energy and even larger data samples than the √ s = 8 TeV sample will be needed to establish this cross-section measurement with higher precision.

Summary and Outlook
In the first two years of the LHC physics program, a new energy regime was investigated with high precision; in some measurements, the experimental systematic uncertainties are at the percent level and the statistical uncertainties are even smaller. A similar precision was reached at previous colliders usually after many years of running. Not only the highest available collision energies and large luminosities provided by the LHC, but also the remarkable performance of the two general purpose detectors, ATLAS and CMS, are the basis of this success.
This article is the first comprehensive review of all major results regarding the production of single heavy gauge bosons at a collision energy of √ s = 7 TeV. Summaries, 13 defined by m ll > 50 GeV, p T, jet > 25 GeV, |η jet | < 4.0, m j j > 120 GeV