Isolating semi-leptonic H → W W ∗ decays for Bell inequality tests

We present a method for identifying H → WW ∗ → (cid:96)νjj in the presence of large Standard Model backgrounds and illustrate how this decay mode may be applied to the study of Bell-type Inequalities. Our ﬁndings reveal the feasibility of complete reconstruction of such Higgs decays and the eﬃcacy of our suggested combination of selection criteria in eﬀectively mitigating the otherwise overwhelming backgrounds. Our approach is based on a combination of bottom and charm tagging, alongside innovative reconstruction techniques. A realistic simulation based on publicly available object identi-ﬁcation, reconstruction, and tagging eﬃciencies from the ATLAS experiment is used to explore the potential sensitivity to violations of the Collins-Gisin-Linden-Massar-Popescu (CGLMP) inequality in existing and expected future data collected at the Large Hadron Collider (LHC). It is found that, the proposed method provides a powerful means of distinguishing the Higgs decay mode from the background, allowing us to achieve an expectation of approximately 3 σ signiﬁcance in detecting violations of these inequalities with 300 fb − 1 of data, soon-to-be collected by LHC.


Introduction
The production of Higgs bosons decaying to a pair of W bosons has been observed by the ATLAS and CMS collaborations at √ s = 8 TeV, and more recently has been probed at √ s = 13 TeV by both collaborations.The total cross-section is measured to a precision of 12%/12% in the gluongluon fusion production mode and to 25%/39% in the associated vector boson production modes by ATLAS and CMS, respectively [1,2].However, in each of these results, only the fully leptonic decay modes (H → W W * → ℓνℓν, where ℓ = e, µ 1 ) were considered due the significant Standard Model (SM) background processes that mimic this final state when considering cases where one or both W 's may decay hadronically.Due to the relatively low leptonic branching fraction of the W , these results have large statistical uncertainties, at the same level of the experimental and theoretical systematic uncertainties.Combining the currents results with events collected from final states not restricted to be fully leptonic would more than double the statistics available for the H → W W * measurements.Another severe limitation of the dileptonic final states is the presence of two neutrinos, that makes high resolution reconstruction of the Higgs boson kinematics almost impossible.This limits the potential to measure phenomena unique to the Higgs boson with LHC data, such as a proposed measurement of quantum entanglement in a qutrit system [3].In this article, we propose a method to identify and fully reconstruct the more prevalent semi-leptonic decay mode (H → W W * → ℓνjj, where j = q = c, s) whilst also reducing the SM background to a manageable level.Furthermore, we demonstrate the utility of these techniques by investigating the potential to measure Bell Inequality violation in Higgs boson decays using a combination of data already collected at the LHC experiments during Run2 ( √ s = 13 TeV) and data that are being collected in the emergent Run3 ( √ s = 13.6 TeV) and the future High Luminosity LHC (HL-LHC).

Background Suppression and Event Reconstruction
In the H → W W * decay mode, one of the W bosons is necessarily off its mass-shell (indicated by an asterisk) due to the mass of the Higgs boson (∼ 125 GeV) being significantly lower than twice the on-shell W mass (∼ 160 GeV).All studies of the H → W W * process have so-far focused on the fully leptonic final state (ℓνℓν) since this is the only decay mode that is not overwhelmed by background events from other SM processes.In particular, the semi-leptonic decay mode (H → W W * → ℓνqq ′ ) is not normally considered as it is dwarfed by the W +jets process.The crosssection of semi-leptonic H → W W * production is 4.3 pb (m H = 125.0GeV) [4,5] but the crosssection of the W (→ lν)+jets process is ∼ 42000 pb at next-to-leading-order accuracy in perturbative QCD [6].In order to successfully isolate the Higgs signal, the W +jets process must be suppressed by 4 orders of magnitude.
The study of this specific final state offers advantages beyond the mere addition of a new channel to Higgs measurements, as it facilitates the full reconstruction of the Higgs system, a task that proves challenging in fully-leptonic H → W W * decays.
In hadron collider experiments, neutrinos are not measured directly but their presence is inferred by requiring momentum conservation in the plane transverse to the beam direction.It is possible to determine the total p x and p y components of the sum of all neutrinos present in the event via missing transverse momentum (P miss x/y ), but not their p z or, in the case of multiple neutrinos, their individual p x or p y components.This presents a problem for the measurement of the properties of the Higgs boson, such as the measurement of Bell inequalities proposed by Barr [3], because these require that the Higgs itself, and all of its decay particles, be fully reconstructed.At the time of writing, no dedicated studies exist on how to solve this problem for the fully-leptonic H → W W * case.In the semi-leptonic case, if the leptonically-decaying W were assumed to be on-shell (and therefore, the hadronic W to be off-shell), the on-shell W mass could be used as an additional constraint, along with the P miss x/y , to reconstruct the leptonic W . Two additional hadronic jets in an event would then need to be paired under some criterion (such as m(jj) < 80 GeV) to reconstruct the off-shell hadronic W and both would need to combine to reconstruct a Higgs boson with a mass of 125 GeV.This approach would be similar to the techniques employed in the reconstruction of semi-leptonic t t decays [7].However, whilst this approach would have an high efficiency on the H → W W * signal, it would also select a large fraction of W +jets, which also contain an on-shell leptonic W and could easily have additional hadronic jets that satisfy whatever criterion were chosen for the Higgs reconstruction.The problem of isolating the signal Higgs with respect to the W +jets background would remain.
We propose instead to sacrifice using the P miss x/y directly and the requirement that the leptonic W be on-shell and impose a stricter requirement on the other side of the event, i.e. that the hadronically decaying W boson be on-shell.We then use a technique adapted from t t dileptonic final state reconstruction called Neutrino Weighting [8] to reconstruct the final state and, additionally, to also suppress the SM W +jets background.

Monte Carlo Processes and Detector Effects
To determine the feasibility of differentiating signal H → W W * events from background SM processes and of observing Bell's inequality violation in Higgs events under real experimental conditions, signal and background events are simulated using the Powheg generator [9,10,11,12].One million signal events are generated for semi-leptonically decaying H → W W * events via the gluon-gluon fusion production mode [13].
Similarly, one million background events each are generated for t t , W +jets and tW processes.In all cases, Pythia [14,15] is used to perform the parton shower and hadronisation of the events.The only exception are the Diboson processes, generated using Sherpa [16].Only processes that generate a single lepton in the final state are considered.
RIVET [17] is used to process the HEPMC [18,19] events and reconstruct physics objects.Physics objects are constructed from stable particles in the event record (particles with a mean lifetime greater than 30 ps).Electrons and muons are identified as originating from a W boson decay (including via intermediate τ leptons) and all photon radiation within a cone of ∆R < 0.1 (where ∆R2 = ∆η 2 + ∆ϕ 2 ) is summed into the lepton four-momentum and removed from the event record (excluding photons from hadrons decays).Jets are reconstructed using the anti-k t algorithm with a distance parameter of R = 0.4 and are tagged as having b or c flavour via ghost matching [20].P miss x/y is defined using only neutrinos originating from a W boson decay (including via intermediate τ leptons).Neutrinos and leptons deriving from hadron decays are included in the input to the jet clustering.These object definitions follow those typically used by the ATLAS and CMS collaborations during LHC Run2.
In order to simulate the effect of the detector response, publicly available acceptance and efficiency values from the ATLAS collaboration are applied to the physics objects.The trigger efficiencies for electrons [21] and muons [22] are based on the assumption of using single lepton unprescaled triggers (with a p T threshold of 20 GeV).The major effect of these efficiencies is to effectively set the minimum p T requirement for the leptons employed in the analysis.The lepton identification and reconstruction efficiencies for electrons and muons are taken to be 90% and 96.1%, respectively, based on the medium definitions used by ATLAS [23,24].We assume a jet reconstruction efficiency of unity and neglect the effects due to limited resolution.Jet smearing based on [25] was tested and was found to have a limited impact on the results described in the following sections.Furthermore, recent experimental results that demonstrate the significant potential for improving jet resolution and response through the utilization of machine learning techniques [26] imply that these effects will be subdominant in analyses on real data.Moreover, in real-experiment analyses, the impact of jet resolution and response can be further mitigated by implementing dedicated regions and observables designed to constrain the jet calibration in-situ.

Neutrino Weighting
Neutrino Weighting (NW) is a technique used to reconstruct leptonically decaying W bosons.It has been used to great effect in the dileptonic t t signature, most recently in measurements of spin correlation in t t events by ATLAS [27,28], but it requires some adaptation to be applicable to the H → W W * → ℓνjj signature.NW solves the problem of reconstructing W bosons that decay leptonically by introducing assumptions on the unconstrained parameters.In the case of dileptonic t t, these assumptions are the values for the pseudo-rapidity of each of the two neutrinos in the final state.In our modification, these assumptions are the pseudo-rapidity of the neutrino (η ν ) and the off-shell mass of the leptonically decaying W boson (m W (lep) ).Since neither of these values are known a-priori, the full system must be reconstructed under many possible assumptions for m W (lep) and η ν .With these two assumptions, as well as an assumption on the mass of the Higgs (which we take to always be 125 GeV) and the reconstructed mass of the hadronically decaying W , the system is fully constrained and can be solved for the neutrino's momentum x and y components 2 .There are zero, one, or two potential real solutions (corresponding to the roots of the polynomial) for each assumption of m W (lep) and η ν .For each assumption, a weight is generated based on the difference between the observed P miss x/y and the reconstructed neutrino kinematics: (1) where ν x/y is the x/y component of the reconstructed neutrino's momentum, P miss x/y is the observed P miss x/y in the experiment in the x/y direction, and σ x/y is the experimental resolution of P miss x/y (the values for these resolutions only serve to scale the weight and have no physical impact on the result if taken equal in x and y).The resultant weight is high when the reconstructed neutrino matches the observed P miss x/y and low when it does not and reaches a maximum when the chosen values for η ν and m W are closest to the true ones.An example of the distribution of weights for a single example collision event is shown in Fig. 1.As can be seen from Eq. 1, the weight is defined between 0 and 1, and the distribution peaks at the same values as the true m W (lep) and η ν .One can also observe a second peak, unrelated to the true values of m W (lep) and η ν corresponding to the second root.In the example event shown in Fig. 1, the solution with the highest weight lies at the same values as the true quantities.A discussion of the performance of the NW algorithm is reported in Section 2.4.
The weight from NW functions is more than just a method of selecting the most-likely values of m W (lep) and η ν , it can also be used to suppress background processes with remarkable efficiency.This is discussed further in Section 2.4.

Charm Tagging
Thanks to the enormous statistics provided by the LHC, it is possible to restrict the decay signature to events where the hadronic W boson decays to a charm and strange3 quark.As we have already assumed one on-shell W bosons decaying hadronically to perform the NW reconstruction, the major backgrounds to the H → W W * signal arise from the production of a single leptonically decaying W boson with additional jets that arise from gluon radiation, from pairs of W bosons where only one boson decays hadronically and from t t events, that contain two W bosons originating by the top quark decay.
Both the ATLAS and CMS experiments heavily utilise sophisticated algorithms to identify (tag) hadronic jets containing B-hadrons, which experiments use to infer the presence of b quarks in an event.The predominant feature used by these algorithms is the long lifetime of B-hadrons, leading to so called displaced vertices in the experiment inner trackers, a feature conspicuously absent from jets originating from so-called light jets (i.e.those initiated by gluons or u, d, s quarks).Jets originating from charm quarks occupy an interesting grey area as hadrons with charm flavour also have sufficiently long lifetimes to create displaced vertices and both ATLAS and CMS have created dedicated c tagging algorithms to identify them.This can be useful for reducing the W +jets background to our signal.The additional jets are more likely to be light jets from gluons or the light flavour quarks (such as uū or d d) compared to heavier quarks pairs (such as cc), whereas in our signal, the on-shell  [29,30].The numbers refer to the probability for the tagging algorithms to assign the c or b flavour to jets containing a B-hadron (b-jet), a c-hadron and no B-hadrons (c-jet) or no heavy-flavour hadrons (l-jet).These numbers are directly used in the analysis presented in this paper to emulate realistic c-tagging performances.hadronic W decays to a ud pair or a cs pair with similar branching fractions.By requiring the presence of exactly one c-tagged jet, the SM background (which contains either none or two) can be reduced.
In this study, we simulate charm tagging using a technique called 'ghost-matching', where jets in Monte Carlo events are determined to have either b or c flavour by setting unstable B and D hadrons energies to zero and allowing them to be clustered into stable particle jets.Any jet with a clustered B hadron is determined to be a b-jet and any with a clustered D hadron and no B hadrons is tagged as a charm jet.Only hadrons with p T above 5 GeV are considered.We then apply efficiency and mistag factors based on recent ATLAS values that are summarised in Table 1 [29,30].
For the H → W W * → cjℓν we are particularly interested in identifying the following features: • Exactly one c-tagged jet.
• One or more light jets (additional light jets could be radiated in the event).• Exactly zero b-tagged jets.
The first two points are derived from the condition that one of the W bosons decays hadronically to a cs pair.The third point is necessary to reject the t t background process, which contains two b quarks in the final state.The second point requires the least consideration; all algorithms employed by LHC experiments are very efficient at rejecting light jets, on the other hand the interplay between b-tagging and c-tagging is non-trivial.The efficiency to mis-tag a b-jet for a c-jet, and vice-versa, can be relevant, as shown in Table 1.

Event Selection
Based on the discussions in Sections 2.3, the following criteria are used to select events: • pre-selection: - If there are multiple combinations of light-jets and c-jet, the pair with mass closest to the m W = 80.6 GeV is chosen as reconstructed hadronic W and employed in the NW.Using this selection we consider two charm tagging possibilities; an idealised case (with perfect identification efficiency and perfect b and light jet rejection) and a realistic case with ϵ c = 40%, shown in Table 1.The latter is representative of the possibility with state-of-theart charm tagging, whereas the former is provided as an indication of the best-case-scenario if significant improvements could be made.In all of the following tables and figures the realistic charm tagging case is used unless otherwise specified.
Given that the mass of the hadronic W is an ingredient of the NW, the charm tagging also slightly influences the performance of the NW.The algorithm has indeed a 78% probability of finding a solution in signal events in the idealised case, while the probability is 72% in the realistic case.The resolution of m W mass reconstruction is slightly below 10% in the idealised case and slightly above in the realistic one.The performance of the NW can be further optimised by fine tuning the number of points scanned in the η ν and m W plane.
The expected fraction of the signal (S) compared to the full simulation, including backgrounds (B), as a function of the weight from NW is presented in Fig. 2.Here the NW requirement is applied on top of the pre-selection and c-tagging Table 2 The expected yields for an integrated luminosity of 300 fb −1 in a selection with idealised charm tagging and Run2-like charm tagging.The signal over the signal plus background (S/(S+B)) is also shown.The uncertainties are statistical only.requirements.Applying only the pre-selection, the fraction of signal over the entire simulated sample is ∼0.0001.The c-tagging requirements help in selecting the signal, increasing the S S+B by a factor ten, but it is only the constraint on the NW score that allows to really isolate the signal.The requirement at 0.7 on the NW score has indeed a 0.45 efficiency on the signal, against a 0.005 efficiency on the backgrounds.On top of the NW score request two additional selection requirements are applied, mainly targeting the reduction of t t background.To enter the selection the two b-jets included in the t t process must be misreconstructed as light jets or c-tagged jets.The limit on the number of jets remove events in the first category while selecting on m(ℓc) removes events from the second category.The m(ℓc) distribution is indeed related to the mass of the parent particle, so it shows a very distinct distribution in the t t case, where the lepton and the c-tagged jet (mis-reconstructed b-jet) derive from a top-quark decay.The expected composition of the data after the pre-selection and c-tagging requirements and after the whole selection is presented in Fig. 3.Even with the requirements on the hadronic W mass and the veto on b-tagged jets, the W +jets and t t processes significantly dominate over the signal before the requirement on the weight from NW is applied.After the whole selection, the contribution from t t and the W +jets processes is drastically reduced and the signal is well visible.

Process
The expected number of signal and background events for 300 fb −1 of √ s = 13 TeV data (the total expected Run2 + Run3 luminosities for ATLAS and CMS) are presented in Table 2. 2843 signal events are expected under realistic charmtagging conditions, with a signal purity of 18%.This is a remarkably high purity given that the cross-section for the background process (W +jets in particular) are orders of magnitude higher than the signal process.Under perfect charm-tagging performance, these numbers increase to 5967 signal events and a purity of 27%.Though such performance is unlikely to be achievable, these values serve to act as an upper bound on the performance of the techniques described in this paper.When including realistic jet smearing the purity degrades to 13% for the realistic charm tagging case and this is likely a fair representation of what is achieve-able under LHC Run2 conditions.However, advanced analysis techniques, such as multivariate approaches, could be employed to maximise the separation between signal and background exploiting other observables that manifest a different behavior between the signal and the main backgrounds, e.g.∆ϕ(ℓ, s) shown in Fig. 3 (bottom).

Observing Bell-Inequality Violation in Higgs Events
Many Bell-type inequalities exist for a variety of different scenarios.For the H → W W * process, which is a pair of three-outcome spin states (qutrit), the Collins-Gisin-Linden-Massar-Popescu (CGLMP) inequality has been shown to be the optimal choice [31,32].The applicability of this inequality to fully-leptonic H → W W * events has already been discussed in previous works.We follow the strategy proposed by Barr [3] with modifications only to account for the semi-leptonic final state.This is possible because the spin analysing power of the down-type quarks deriving by the W decay is 1 as in the charged leptons case.As a consequence the technique introduced in [3] can be applied also to the case presented here using the lepton and the down-type quark instead of the charged lepton pair to reconstruct the observables of interest.The down-type quark employed in the following is identified with the light jet used to reconstruct the hadronic W .
The CGLMP inequality (I 3 ) may be constructed from the expectation values of operators (B) constructed from angular observables based upon a choice of three orthonormal axes (xŷẑ).These axes are defined based on the direction of the W + boson momenta (x) in the W + W − rest frame, the direction perpendicular to the plane formed by x and the direction of the incoming protons (ŷ), and the right-handed remaining orthogonal direction (ẑ) (a more in-depth discussion on these definitions and their similarity to those used in top spin measurements is provided in [3]).The CHLMP inequality is therefore defined as: and this inequality is violated when: The operators are themselves constructed from the products of cosines of spin analysers of the parent W boson (charged leptons and strange-type jets in this topology).For example, B xy CGLMP , following the formalism in [3] is formed from the summation of the averages of three observables: where ξ ± xyz is the cosine of the angle between the spin analyser from the parent W ± boson and the spin analysing axis xŷẑ.Similar sets of three observables exist for the B yz CGLMP and B xz CGLMP operators.In each event the sign of the two W is determined by the sign of the charged lepton.Each of the observables is constructed event-byevent and used to fill histograms from which the means are later determined.The means are then used to construct the relevant operator and, hence, determine the value for the inequality.

Unfolding
In order to correct for detector effects, many collider measurements use a technique called unfolding.We replicate this technique to remove the smearing effects imposed by our identification, efficiency, and reconstruction process using a type of unfolding called Iterative Bayesian Unfolding [33].In particular, we use the procedure described in several ATLAS top quark measurements [34,7] and implemented in the RooUnfold package [35].The nine observables of interest are unfolded using two iterations and extrapolated to the full semi-leptonic phase space, at which point the mean of each observable is extracted to determine ⟨B xy CGLMP ⟩, ⟨B yz CGLMP ⟩, and ⟨B zx CGLMP ⟩.With the choice of axes and selection described in Section 2.4, ⟨B zx CGLMP ⟩ is always the largest and thus is the value taken to determine I xyz 3 .The normalised migration matrices for one of the observables used to build this operator is shown in Fig. 4. The chosen binning results in diagonal matrices and consequently a stable unfolding.The limitations of the reconstruction are also apparent and manifest as a smearing of the reconstructed quantity towards zero.
The unfolded observables employed to determine ⟨B zx CGLMP ⟩ are shown in Fig. 5. Using the event selection, reconstruction, and unfolding described in the previous sections, the expected measurement and sensitivity for I 3 is determined for three different values of integrated luminosity at a centre of mass energy of 13 TeV: 139 fb −1 (corresponding to the amount of data collected by both ATLAS and CMS during LHC Run2), 300 fb −1 (corresponding to the expected data that will be collected by the combination of Run2 and Run3), and 3000 fb −1 (corresponding to the expected total integrated luminosity for HL-LHC).In all cases, the centre of mass energy is 13 TeV though, in practice, the LHC Run3 centre-of-mass energy is slightly higher and the energy for HL-LHC may reach 14 TeV.This small changes in the center of mass energy does not largely effect the numbers projected here, as the signal and background cross-sections are only mildly affected.Moreover it is possible to combine the results on the significance across the different center of mass energies.The uncertainties considered are only statistical, these are evaluated taking into account the unfolding procedure.The statistical uncertainty is estimated using pseudo-experiments, derived smearing every bin of the input distributions with a Poisson distribution.The standard deviation of the unfolded pseudo-experiments in each bin is taken as statistical uncertainty and then propagated to the histograms mean and to I xyz 3 .The expected measured values and significances for each of these luminosity scenarios are presented in Table 3.The central value is 2.45, indicating violation of the I 3 inequality.The expected uncertainties and significances of these values are presented for the charm tagging efficiency described in Section 2.3 and for an idealised case.Though such a case is not realistic for the existing LHC experiments, it provides a useful upper-bound on the results.For 300 fb −1 the expected significance is slightly below 3 sigmas in the realistic case, and above it in the idealised case, suggesting that the evidence could be reached by the LHC experiments already at the end of Run3.In the HL-LHC scenario the expected significance rises to 9 sigmas in the realistic scenario and to 11 in the idealised one.The impact of the assumption of perfect jet reconstruction was tested here using performances derived from ATLAS public jet performance results [25] for the case with non-idealised charm tagging, where the significances are slightly diluted to 1.6, 2.4, and 7.5 sigma for 130 fb −1 , 300 fb −1 ,and 3000 fb −1 , respectively.

Conclusions
In this paper, we have demonstrated a method for isolating a previously inaccessible Higgs decay mode.The adaptation of Neutrino Weighting to semi-leptonic H → W W * final state provides a new tool to separate the signal from the overwhelming background allowing to reach a signal purity of 18%, in combination with c-tagging and b-tagging.We have implemented a simple selection on the score of the NW, but the separation between signal and background could be further enhanced by introducing a multivariate analysis technique that combines the NW with other variables sensitive to the differences between H → W W * and W +jets.Using the NW and the c-tagging combined with the W and Higgs mass constraints, we have fully reconstructed the final state and prototyped a measurement of the CGLMP Bell-type inequality in this topology.The results are promising and a significance of almost 3 sigma is expected with the soon-to-be collected luminosity of 300f b −1 at the LHC, whilst observation could be reached using the data sample expected for HL-LHC.We found that the reconstruction of the off-shell leptonic W boson and the identification of the c-quark (and hence, s-quark) type jets are limiting factors for the measurement.Innovative charm tagging, with increased efficiency and an improved mis-tag rate for bjets would represent a large improvement on the result both in terms of reduction of the statistical uncertainty and in the increase of the signal purity.This kind of improvement seems feasible in the near future, given recent advancements in flavour tagging algorithms [36,37].We focused only on the CGLMP inequalities as a probe to study the quantum nature of the Higgs boson but the same approach could be applied to measure other entanglement witnesses valid for qutrit systems introduced in recent phenomenological studies [38].This could facilitate observation of entanglement in H → W W * events before reaching the statistic required to observe the violation of Bell inequalities.

Fig. 1
Fig.1The distribution of weights for each assumption of leptonic W boson mass and neutrino pseudo-rapidity, with markers indicating the location of the true values of these parameters and the scan point with the highest weight.
Exactly 1 lepton with p T > 20 GeV -Exactly 0 b-tagged jets • c-tagging selection: -2 or more jets, exactly one of which must be c-tagged.-At least 1 (c-jet,l-jet) pair with |m cl − 80.6| < 10 GeV • A reconstructed leptonic W boson from NW with w > 0.7.• Maximum 2 light jets.• Invariant mass of the lepton and the c-tagged jet m(ℓc) < 80 GeV.

Fig. 2
Fig.2The signal over signal plus background as a function of the requirement applied on the score from the Neutrino Weighting.Only events passing the pre-selection and the c-tagging selection are included in the plot.

Fig. 3
Fig. 3 Number of events as a function of the azimuthal angle between the lepton and the down-type quark (∆ϕ(l, s)) applying only the pre-selection and c-tagging requirements (top) and applying the whole selection (bottom).

Fig. 4
Fig. 4 Migration matrices for one of the three observables composing the expectation value of the zx Bell Operator.

Fig. 5
Fig. 5 Unfolded histograms for the three observables composing the expectation value of the zx Bell Operator.The uncertainties on each bin are statistical.

Table 1 c
and b-tagging efficiencies used in the ATLAS experiment

Table 3
The expected sensitivity for different values of integrated luminosity for a realistic and idealised case.The uncertainties are statistical only.