Search for a Higgs boson in the mass range from 145 to 1000 GeV decaying to a pair of W or Z bosons

A search for a heavy Higgs boson in the H to WW and H to ZZ decay channels is reported. The search is based upon proton-proton collision data samples corresponding to an integrated luminosity of up to 5.1 inverse femtobarns at sqrt(s) = 7 TeV and up to 19.7 inverse femtobarns at sqrt(s) = 8 TeV, recorded by the CMS experiment at the CERN LHC. Several final states of the H to WW and H to ZZ decays are analyzed. The combined upper limit at the 95% confidence level on the product of the cross section and branching fraction exclude a Higgs boson with standard model-like couplings and decays in the range 145<m[H]<1000 GeV. We also interpret the results in the context of an electroweak singlet extension of the standard model.

Previous searches at the LHC for heavy SM-like Higgs bosons have been reported by ATLAS and CMS. Based on a dataset of 4.7 fb −1 at √ s = 7 TeV and up to 5.9 fb −1 at √ s = 8 TeV, and combining all channels listed in Ref. [10], ATLAS excludes a SM-like heavy Higgs boson of 131 < m H < 559 GeV at 95% CL [10]. The CMS collaboration reported a search in the WW and ZZ decay channels using an initial dataset of 5.1 fb −1 at √ s = 7 TeV and 5.3 fb −1 at √ s = 8 TeV, searching in the mass range 145 < m H < 1000 GeV, and excluding Higgs boson masses up to 710 GeV at 95% CL [57]. In this paper, we report on an extension of this search using the complete 7 and 8 TeV dataset. In addition, the search is interpreted in the context of the SM expanded by an additional EW singlet. Both the SM-like heavy Higgs boson as well as the EW singlet are denoted as H here.
The analysis uses proton-proton collision data recorded with the CMS detector, corresponding to integrated luminosities of up to 5.1 fb −1 at √ s = 7 TeV and up to 19.7 fb −1 at √ s = 8 TeV. The analysis is performed in a mass range 145 < m H < 1000 GeV exploiting both the H → WW and H → ZZ decay channels, with the lower boundary being chosen to reduce contamination from h(125). In the case of a Higgs boson decaying into a pair of W bosons, the fully leptonic (H → WW → ν ν) and semileptonic (H → WW → νqq) final states are considered in the analysis. For a Higgs boson decaying into two Z bosons, final states containing four charged leptons (H → ZZ → 2 2 ), two charged leptons and two quarks (H → ZZ → 2 2q), and two charged leptons and two neutrinos (H → ZZ → 2 2ν) are considered, where = e or µ and = e, µ, or τ.

CMS detector
The central feature of the CMS apparatus is a superconducting solenoid of 6 m internal diameter, providing a magnetic field of 3.8 T. Within the solenoid volume are a silicon pixel and strip tracker, a lead tungstate crystal electromagnetic calorimeter (ECAL), and a brass and scintillator hadron calorimeter, each composed of a barrel and two endcap sections. Muons are measured in gas-ionization detectors embedded in the steel flux-return yoke outside the solenoid. Extensive forward calorimetry complements the coverage provided by the barrel and endcap detectors. The first level of the CMS trigger system, composed of custom hardware processors, uses information from the calorimeters and the muon detectors to select the most

Signal model and simulations
interesting events in a fixed time interval of less than 4 µs. The high level trigger processor farm further decreases the event rate from around 100 kHz to less than 1 kHz, before data storage. A more detailed description of the detector as well as the definition of the coordinate system and relevant kinematic variables can be found in Ref. [58].

Signal model and simulations
Several Monte Carlo event generators are used to simulate the signal and background event samples. The Higgs boson signal samples from gluon fusion (ggF, gg → H), and vector boson fusion (VBF, qq → qqH), are generated with POWHEG 1.0 [59][60][61][62][63] at next-to-leading order (NLO) and a dedicated program [64] used for angular correlations. Samples of WH, ZH, and ttH events are generated using the leading-order (LO) PYTHIA 6.4 [65] program. At the generator level, events are weighted according to the total cross section σ(pp → H) [66], which contains contributions from ggF computed to next-to-next-to-leading order (NNLO) and next-tonext-to-leading-log (NNLL), and from weak-boson fusion computed at NNLO. The WW(ZZ) invariant mass, m WW (m ZZ ), lineshape is affected by interference between signal and SM background processes. The simulated m H lineshape is therefore corrected to match theoretical predictions [67][68][69] using the complex-pole scheme for the Higgs boson propagator. The procedure for including lineshape corrections and uncertainties from interference of the signal with background processes for both ggF and VBF production are described below in the discussion of the lineshape corrections applied for the EW singlet interpretation.
The background contribution from qq → WW production is generated using MADGRAPH 5.1 [70], and the subdominant gg → WW process is generated at LO with GG2WW 3.1 [71]. The qq → ZZ production process is simulated at NLO with POWHEG, and the gg → ZZ process is simulated at LO using GG2ZZ 3.1 [72]. Other diboson processes (WZ, Zγ ( * ) , Wγ ( * ) ), Z+jets, and W+jets are generated with PYTHIA and MADGRAPH. The tt and tW events are generated at NLO with POWHEG. For all samples, PYTHIA is used for parton showering, hadronization, and underlying event simulation. For LO generators, the default set of parton distribution functions (PDF) used to produce these samples is CTEQ6L [73], while CT10 [74] is used for NLO generators. The tau lepton decays are simulated with TAUOLA [75]. The detector response is simulated using a detailed description of the CMS detector, based on the GEANT4 package [76], with event reconstruction performed identically to that of recorded data. The simulated samples include the effect of multiple pp interactions per bunch crossing (pileup). The PYTHIA parameters for the underlying events and pileup interactions are set to the Z2 (Z2 * ) tune [77] for the 7 (8) TeV simulated data sample, with the pileup multiplicity distribution matching the one observed in data.
The data are analysed to search for both a beyond the standard model (BSM) case in the form of an EW singlet scalar mixed with the recently discovered Higgs boson, h(125), as well as a heavy Higgs boson with SM-like couplings. The couplings of the two gauge eigenstates (h(125) and EW singlet) are phenomenologically constrained by unitarity and the coupling strength of the light Higgs boson is therefore reduced with respect to the SM case. The unitarity constraint is ensured by enforcing C 2 + C 2 = 1, where C and C are defined as the scale factors of the couplings with respect to the SM of the low-and high-mass Higgs boson, respectively. The EW singlet production cross section is also modified by a factor µ and the modified width is Γ ; they are defined as where B new is the branching fraction of the EW singlet to non-SM decay modes. An upper limit at 95% CL can be set indirectly as C 2 < 0.28 using the signal strength fits to the h(125) boson as obtained in Ref. [78].
This paper focuses on the case where C 2 ≤ (1 − B new ). In this regime the new state is expected to have an equal or narrower width with respect to the SM case. Results are presented distinguishing between the B new = 0 and B new > 0 cases. Under this hypothesis, signal samples with different Higgs boson widths are generated, scanning the C 2 and B new space. We follow the recommendations of the "LHC Higgs Cross Section Working Group" [66] described below.
The SM signal mass lineshape generated with POWHEG is weighted in order to simulate the narrow scalar singlet lineshape. For the ggF production mode, the weights are calculated using either the GG2ZZ generator for the ZZ channel, or the POWHEG and MCFM 6.2 [79] generators for the interference calculation for the WW channel. For the VBF production mode, the interference weights are computed using the PHANTOM 1.2 [80] generator, where the signal-only lineshape at LO is weighted based on results obtained with MADGRAPH generator predictions. The weights are defined as the ratio of the sum of a narrow resonance signal plus interference and the standard model signal lineshape as generated. The contribution from the interference term between the BSM Higgs boson and the background is furthermore assumed to scale according to the modified coupling of the Higgs boson as (µ + I) BSM = µ SM C 2 + I SM C , where µ(I) is the signal strength (interference) in the BSM or SM cases. This assumption is based on the hypothesis that the couplings are similar to the SM case and simply rescale due to unitarity constraints. Systematic uncertainties considered for this procedure are detailed later.
If the new resonance has a very small width, its production will tend to interfere less with the background continuum. Thus in the most interesting region of low-C 2 , the effect of the interference and its exact modeling is of limited importance. Any possible interference between h(125) and its EW singlet partner [81,82] is assumed to be covered by a conservative systematic uncertainty. In addition to the EW singlet, the analysis searches for a heavy Higgs boson that gets produced and decays like the SM Higgs boson, but has a higher mass and interferes with h(125).

Event reconstruction
CMS uses a particle-flow (PF) reconstruction algorithm [83,84] to provide an event description in the form of particle candidates, which are then used to build higher-level objects, such as jets and missing transverse energy, as well as lepton isolation quantities. Not all the channels considered here use the same selection criteria for their objects, but the common reconstruction methods are listed below.
The high instantaneous luminosity delivered by the LHC provides an average of about 9 (21) pileup interactions per bunch crossing in 7 (8) TeV data, leading to events with several possible primary vertices. The vertex with largest value of the sum of the square of the transverse momenta (p T ) for the associated tracks is chosen to be the reference vertex. According to simulation, this requirement provides the correct primary vertex in more than 99% of both signal and background events.
Muon candidates are reconstructed by using one of two algorithms: one in which tracks in the silicon tracker are matched to hits in the muon detectors, and another in which a combined fit is performed to signals in both the silicon tracker and the muon system. Other identification criteria based on the number of measurements in the tracker and in the muon system, the fit quality of the muon track, and its consistency with its origin from the primary vertex are also imposed on the muon candidates to reduce the misidentification rate [85]. In some of the channels, vetos are imposed on additional low-p T muons in the event, whose tracks in the silicon tracker fulfill more stringent requirements, but whose primary vertex association can be more relaxed. We call these muons soft muons.
Electron candidates are reconstructed from superclusters, which are arrays of energy clusters along the φ direction in the ECAL, matched to tracks in the silicon tracker. A complementary algorithm reconstructs electron candidates by extrapolating measurements in the innermost tracker layers outward to the ECAL. Trajectories from both algorithms are reconstructed using the Gaussian sum filter algorithm [86], which accounts for the electron energy loss by bremsstrahlung. Additional requirements are applied to reject electrons originating from photon conversions in the tracker material. Electron identification relies on a multivariate discriminant that combines observables sensitive to the bremsstrahlung along the electron trajectory, the geometrical and momentum-energy matching between the electron trajectory and the associated supercluster, as well as ECAL shower shape observables. Electron candidates with an ECAL supercluster in the transition region between ECAL barrel and endcap (1.4442 < |η| < 1.5660) are rejected, because the reconstruction of electrons in this region is compromised [87].
The PF candidates are used to reconstruct hadronic tau candidates, τ h , with the "hadron plus strip" algorithm [88] that is designed to optimize the performance of τ h identification and reconstruction by considering specific τ h decay modes. The neutrinos produced in all τ decays escape detection and are ignored in the τ h reconstruction. The algorithm provides high τ h identification efficiency, approximately 50% for the range of τ h energies relevant for this analysis, while keeping the misidentification rate for jets at the level of ∼ 1%.
Leptons produced in the decay of W and Z bosons are expected to be isolated from hadronic activity in the event. Isolation is defined as the scalar sum of the transverse momenta of the PF candidates found (excluding the selected leptons in the event themselves) in a cone of radius R = √ ∆η 2 + ∆φ 2 = 0.4, with φ being measured in radians, built around each lepton. We require that the isolation sum is smaller than 20% (15%) of the muon (electron) transverse momentum. To account for the contribution to the isolation sum from pileup interactions and the underlying event, a median energy density is determined on an event-by-event basis using the method described in Ref. [89]. For electron candidates, an effective area that is proportional to the isolation cone, is derived to renormalize this density estimate to the number of pileup interactions, and is subtracted from the isolation sum. For muon and tau candidates, the correction is done by subtracting half the sum of the transverse momenta of the charged particles not associated to the primary vertex in the cone of interest. Soft muons do not need to fulfill isolation requirements.
The combined efficiency of lepton reconstruction, identification, and isolation is measured using observed Z decays and ranges between 90% and 97% for muons, between 70% and 90% for electrons, and approximately 50% for hadronic taus depending on the p T and η of the leptons.
Jets are reconstructed from PF candidates [90] using the anti-k T clustering algorithm [91] with a distance parameter of 0.5 (called AK5 jets), as implemented in the FASTJET package [92], and with the Cambridge-Aachen algorithm [93] with a distance parameter of 0.8 (called CA8 jets); when not otherwise specified we use the term jets to mean AK5 jets throughout this paper.
Any reconstructed jet overlapping with isolated leptons within a radial distance of 0.5 (0.8) for AK5 (CA8) jets is removed in order to avoid double counting the lepton as a jet. AK5 (CA8) jets are required to have |η| < 4.7 (2.4). At hadron level, the jet momentum is defined as the vectorial sum of all particle momenta in the jet, and is found in the simulation to be within 5-10% of the true momentum over the whole p T range and detector acceptance. A correction is applied to the jet p T to take into account the extra energy clustered in jets due to additional proton-proton interactions within the same bunch crossing [89,94]. Other jet energy scale (JES) corrections applied are derived from the simulation, and are calibrated with in situ measurements of the energy balance of dijet and Z/γ+jets events. For some of the channels in this analysis, a combinatorial background arises from low-p T jets from pileup interactions, which get clustered into high-p T jets. At √ s = 8 TeV the number of pileup events is larger than at √ s = 7 TeV, and a multivariate selection is applied to separate jets stemming from the primary interaction and those reconstructed due to energy deposits associated with pileup interactions [95]. The discrimination is based on the differences in the jet shapes, on the relative multiplicity of charged and neutral components, and on the different fraction of transverse momentum, which is carried by the hardest components. Within the tracker acceptance, the tracks belonging to each jet are also required to be compatible with the primary vertex.
At high Higgs boson masses, the p T of the Z or W boson is high enough that the two quarks from the vector boson decay are expected to be reconstructed as a single CA8 jet, or merged jet. To improve background rejection and jet mass resolution, we apply a jet pruning algorithm [96]. Additionally, the "N-subjettiness ratio" variable τ 2 /τ 1 , a measure of the compatibility of a jet having N = 2 subjets, is used to reduce the backgrounds [96,97]. We require τ 2 /τ 1 < 0.5 to avoid contamination from jets originating from the hadronization of gluons and single quarks.
In some of the channels included in this analysis, the identification of jets originating from b quarks is important. These b jets are tagged with dedicated algorithms [98], which are applied either directly to the AK5 jets, or to the subjets of the merged jets. The b-tagging algorithm used is either the Combined Secondary Vertex algorithm, Jet Probability algorithm, or Track Counting High Efficiency algorithm depending on the channel. Tagged b-jet candidates are required to have p T > 30 GeV and to be within the tracker acceptance (|η| < 2.4). For jets in this kinematic range with a b-tagging efficiency of 70%, the misidentification probability from light quark and gluon jets is approximately 1%.
The missing transverse momentum vector, E miss T , is defined as the projection on the plane perpendicular to the beams of the negative vector sum of the momenta of all reconstructed particles in an event [84,99]. Its magnitude is referred to as E miss T .

Data analysis
The results presented in this paper are obtained by combining searches exploiting different Higgs boson production and decay modes as detailed in Table 1. All final states are exclusive, without overlap between channels. For the WW → ν ν and ZZ → 2 2 channels a mass range starting from 110 GeV has been analyzed in other contexts (e.g. [12]), but in this paper both analyses are restricted to searches above 145 GeV. In the rest of this section, the individual analysis strategies and details are defined including a discussion of the leading systematic uncertainties. In the next section, statistical interpretations of the individual and combined searches are given for the Higgs boson with SM-like couplings hypothesis as well as the EW singlet extension of the SM. Table 1: A summary of the analyses included in this paper. The column "H production" indicates the production mechanism targeted by an analysis; it does not imply 100% purity in the selected sample. The main contribution in the untagged categories is always ggF. The (jj) VBF refers to a dijet pair consistent with the VBF topology. (jj) W(Z) and (J) W(Z) refer to a dijet pair and single merged jet from a Lorentz-boosted W (Z) with an invariant mass consistent with a W (Z) dijet decay, respectively. The superscript "0,1,2 b tags" refers to the three possible categories of b tag multiplicities. Exclusive final states are selected according to the lepton and reconstructed jet content of the event. The mass range under investigation and the mass resolution are also listed. Mass ranges differ with the sensitivities of each channel.

H → WW → ν ν
In the H → WW → ν ν channel the Higgs boson decays to two W bosons, both of which decay leptonically, resulting in a signature with two isolated, oppositely charged, high-p T leptons (muons or electrons) and large E miss T due to the undetected neutrinos. A complete description of the analysis strategy is given in Ref. [18]. For this analysis, we require triggers with either one or two high-p T muons or electrons. The single muon or electron triggers are based on relatively tight lepton identification with p T thresholds from 17 to 25 GeV (17 to 27 GeV) in the muon (electron) channel. The higher thresholds are used for periods of higher instantaneous luminosity. The dilepton trigger p T thresholds for the leading and trailing leptons were 17 and 8 GeV, respectively.
Candidate events must contain two reconstructed leptons with opposite charge, p T > 20 GeV for the leading lepton, and p T > 10 GeV for the subleading one. Only muons (electrons) with |η| < 2.4 (2.5) are considered in this channel. The analysis is very similar to that in the Higgs boson discovery [11,12], but additionally uses an improved Higgs boson mass lineshape model, and a multivariate shape analysis (MVA) [100] for data taken at √ s = 8 TeV.
Events are classified into three mutually exclusive categories, according to the number of reconstructed jets with p T > 30 GeV. The categories are characterized by different signal yields and signal-to-background ratios. In the following, these are referred to as 0-jet, 1-jet, and 2-jet multiplicity categories. Events with more than two jets are considered only if they are con-sistent with the VBF signature: a dijet mass of the two highest p T jets greater than 500 GeV, |∆η jj | > 3.5, and no additional jets with p T > 30 GeV in the η region between these two leading jets. Signal candidates are further divided into same-flavor (SF) dilepton (µ + µ − or e + e − ) and different-flavor (DF) dilepton (µ ± e ∓ ) categories. The bulk of the signal arises through direct W boson decays to muons or electrons, with the small contribution from W → τν → + X decays implicitly included.
In addition to high-p T isolated leptons and minimal jet activity, significant E miss T is expected to be present in signal events, as opposed to none to moderate E miss T in background. For this channel, an E miss T, pr variable is employed, defined as (i) the magnitude of the E miss otherwise. This observable more efficiently distinguishes Z/γ * → τ + τ − background events in which the E miss T is preferentially aligned with the leptons, and Z/γ * → + − events with mismeasured E miss T . Since the E miss T, pr resolution is degraded as pileup increases, the minimum of two different observables is used: the first includes all particle candidates in the event, while the second uses only the charged particle candidates associated with the primary vertex. Events with E miss T, pr > 20 GeV are selected for this analysis. The backgrounds are suppressed using techniques described in Refs. [11,12]. Top quark background is controlled with a selection based on the presence of a soft muon and b-jet tagging [98]. Rejection of events with a third lepton passing the same requirements as the two selected leptons reduces both WZ and Wγ * backgrounds.
The Drell-Yan (DY) process produces SF lepton pairs (µ + µ − and e + e − ) and therefore additional requirements are applied for the SF final state. Firstly, the resonant component of the DY background is rejected by requiring a dilepton mass outside a 30 GeV window centered on the Z boson mass. The remaining off-peak contribution is further suppressed by requiring significant missing transverse energy in the event. For the √ s = 7 TeV data, we require E miss T, pr > (37 + N vtx /2) GeV directly, while for √ s = 8 TeV data, a boosted decision tree (BDT) multivariate discriminant including E miss T, pr is used. For events with two jets, the dominant source of misreconstructed E miss T is the mismeasurement of the hadronic recoil, and optimal performance is obtained by requiring E miss T > 45 GeV. Finally, the momenta of the dilepton system and of the most energetic jet must not be back-to-back in the transverse plane. For the √ s = 7 TeV data, we require the angle to be less than 165 degrees , while for √ s = 8 TeV data, this information is included in the BDT. These selections reduce the DY background by three orders of magnitude, while rejecting less than 50% of the signal.
The final analysis in the DF and in the SF 0-jet or 1-jet categories is performed using a twodimensional fit in two observables, the dilepton invariant mass m and the transverse mass m T determined between the transverse dilepton system, p T , and the E miss where ∆Φ( p T , E miss T ) is the azimuthal angle between the dilepton transverse momentum and the E miss T . Figure  Observed Observed For the VBF production mode [101][102][103][104], the cross section is roughly ten times smaller than for ggF at lower m H hypotheses and is roughly three times smaller at the highest m H hypothesis. We optimize the selection in the 2-jet category to tag these VBF-type events by requiring the mass of the dijet system to fulfill m jj > 500 GeV, and the angular separation of the two jets to pass |∆η jj | > 3.5. Given the small event yield in this category, the signal extraction in the DF category is performed using a one-dimensional fit in m where an m H dependent requirement on the transverse mass is imposed. A counting analysis is performed in the SF category and is used as a cross-check in the DF category.
The normalization of the background contributions relies on observed events rather than simulation whenever possible, and exploits a combination of techniques [11,12]. The tt background is estimated by extrapolation from the observed number of events with the b-tagging requirement inverted. The DY background measurement is based on extrapolation from the observed number of µ + µ − and e + e − events with the Z boson veto requirement inverted. The background of W + jets and Quantum Chromodynamics (QCD) multijet events is estimated by measuring the number of events with one loosely isolated lepton. The probability for such loosely isolated nongenuine leptons to pass the tight isolation criteria is measured in observed data using multijet events. In the 0-jet and 1-jet bins, the nonresonant WW contribution is estimated from a fit to the data while in the 2-jet bin it is estimated from simulation. Other backgrounds, such as V+Z/γ * and triple boson production (VVV) are estimated from simulation and are small.
Experimental effects, theoretical predictions, and the choice of event generators (POWHEG, GG2WW, MADGRAPH, PHANTOM) are considered as sources of systematic uncertainty, and their impact on the signal efficiency is assessed. The overall signal normalization uncertainty is estimated to be about 20%, and is dominated by the theoretical uncertainty associated with missing higher-order QCD corrections and PDF uncertainties, estimated following the PDF4LHC recommendations [74,[105][106][107][108]. The total uncertainty in the background estimation in the H → WW signal region is about 15% and is dominated by the statistical uncertainty in the observed number of events in the background control regions.

Unmerged-jet category
In the H → WW → νqq channel we search for a Higgs boson decaying to WW, where one W decays leptonically, thus providing a trigger handle for the event, while the other decays hadronically. This channel has a larger branching fraction than the two-lepton final state and allows one to reconstruct the Higgs boson candidate invariant mass [109]. The final state consists of an isolated electron or muon, E miss T , and two separated jets. The main experimental challenge is to control the large W+jets background.
We use data collected with a suite of single-lepton triggers mostly using p T thresholds of 24 (27) GeV for muons (electrons). Basic kinematic selections are applied to the final-state objects to reduce the background contribution. The muon (electron) is required to be isolated and have p T > 25 (35)  (1 − cos ∆Φ), with ∆Φ being the azimuthal angle between the lepton p T and E miss T , needs to be greater than 30 GeV. The two leading-p T jets in the event each must have p T > 30 GeV and must together have an invariant mass, m jj , between 66 and 98 GeV. An unbinned maximum likelihood fit is performed on m jj to determine the background normalizations in the signal region for each Higgs mass hypothesis independently. Events that contain a b-tagged jet are vetoed in order to reduce the background from top quark decays.
To further exploit the differences in kinematics between the signal and background, a likelihood discriminant is constructed using angles between the Higgs boson decay products that fully describe the Higgs boson production kinematics [19,64] and the lepton charge. The lepton charge provides discrimination power because of the charge asymmetry in the background, which is not present in the signal. This approach improves the expected sensitivity to a Higgs boson across the entire mass range.
The background normalization in the signal region is extracted for each Higgs boson mass hypothesis independently with an unbinned maximum likelihood fit to the dijet invariant mass distribution, m jj , of the two leading jets. The signal region corresponding to the W boson mass window, 66 < m jj < 98 GeV, is excluded from the fit. The background is overwhelmingly due to events from W+jets production. The normalization of the W+jets component is a free parameter and is determined in the fit. The electroweak diboson, tt, and single top shapes are based on simulation and their normalizations are constrained to the theoretical predictions with associated uncertainties. In the case of diboson backgrounds, both WW and WZ normalizations come from NLO predictions. The uncertainties in the normalization of the W+jets component obtained from the fit, as well as those from all other backgrounds, are included in the limit calculation as systematic uncertainties.
For the signal search, we use the binned distribution of m WW , which is computed using a neutrino longitudinal momentum, p z , that is determined from the constraint that the leptonically decaying W boson is produced on-shell. The ambiguity of two possible solutions is resolved by taking the solution that yields the smaller |p z | value for the neutrino. According to simulation over 85% of signal events are assigned the correct p z value, thus improving the mass resolution, especially at low Higgs mass. The contribution of each background in the signal region (66 < m jj < 98 GeV) is described by an expected shape and normalization. The distri-bution of the W+jets contribution is parameterized with a polynomial functional form determined through simulation. The parameters of the function are determined from the observed data spectrum in the binned likelihood fit, which is also used to determine the exclusion limits. The other background shape distributions are parameterized using simulation, and their normalizations and uncertainties are included as nuisance parameters. Figure 2 depicts the m WW distribution for the muon category for two different Higgs boson mass hypotheses. The observed data are compared to background and signal expectations.  Experimental effects, theoretical predictions, and uncertainties due to the choice of fit functions are considered as sources of systematic uncertainty. Because of the large background, the dominant source of systematic uncertainty is the shape uncertainty of the W+jets m WW distribution, followed by the normalization uncertainty that is extracted from the m jj fit. The main uncertainty in the signal normalization stems from the uncertainty in the efficiency of the likelihood discriminant selection.

Merged-jet category
In the highest mass range of this search, from 600 to 1000 GeV, the p T of the decaying W bosons is typically greater than 200 GeV. At this p T , the daughter quarks of the hadronically decaying W are often merged into a single jet to the point where traditional dijet searches cannot be performed. For a signal mass of 600 GeV (1 TeV), and signal events falling in the detector acceptance, approximately 65% (82%) of the hadronic W decay products are contained in a cone of ∆R < 0.8; alternatively, approximately 10% (42%) of the hadronic W decay products are separated by a distance of ∆R < 0.5 and would not be reconstructed by the standard CMS AK5 jet finding algorithm. The larger cone CA8 jet affords more signal acceptance in the single jet signature while not losing events when the decay products are separated by ∆R < 0.5. In this case, we use jet substructure techniques for identifying jets that have originated from a hadronically decaying W boson with high Lorentz boost.
As in the unmerged case, we use data collected with the single electron or muon trigger. The dominant background is W+jets with a smaller background contribution from tt. Remaining backgrounds arise from WW, WZ, ZZ, and single top quark production.
The hadronic W boson candidate is reconstructed with CA8 jets to increase acceptance. To reduce the contribution of quark-and gluon-initiated jets from QCD processes, a selection is made on the pruned jet mass of the CA8 jet of 65 < m J < 105 GeV and the N-subjettiness ratio τ 2 /τ 1 < 0.5.
The kinematic selections of muons and electrons are slightly more stringent than those for the unmerged case, with the p T thresholds of isolated muons and electrons being 30 and 35 GeV, respectively. The E miss T requirement is increased to be greater than 50 (70) GeV in the muon (electron) channel. The p T of both, leptonically and hadronically decaying W boson candidates, is required to be greater than 200 GeV in order to select events with a large Lorentz boost. Additional selections are made to ensure that the W boson candidates are sufficiently separated in a back-to-back topology. As in the unmerged category, there is an additional requirement to veto events with a b tagged jet in order to reduce the background from top quark decays.
To increase the sensitivity of the analysis, the analysis is split into exclusive jet multiplicity categories: 0+1-jet and 2-jets. These additional AK5 jets must pass a p T threshold of 30 GeV and be separated from the hadronic W candidate in the event by a distance ∆R > 0.8. Additionally, the 2-jet category has further requirements to better identify a topology consistent with the VBF production mode. The two highest p T AK5 jets in the event are required to satisfy |∆η jj | > 2.5 and m jj > 250 GeV. There is also a requirement on the invariant mass of the W boson candidates and the nearest AK5 jet to be greater than 200 GeV in order to reject other top quark background. The final discriminating distribution is the m WW shape reconstructed from the lepton + E miss T + CA8 jet system. Because the number of events in the 2-jet category is limited, the muon and electron datasets are merged.
The background is estimated using observed data rather than simulation wherever possible. The dominant W+jets background normalization is determined from a fit to the pruned jet mass sideband where other backgrounds contribute less significantly. The W+jets shape is determined from extrapolating the pruned jet mass sideband region into the signal region. The tt background is estimated by inverting the veto on AK5 b-tagged jets, which yields a highpurity tt control sample. The normalization of tt in the signal region is then corrected by a data-to-MC scale factor determined from the tt control sample. Finally the tt control sample is also a good source of W bosons with high Lorentz boost, which are used to calibrate the jet substructure selection efficiency. Sources of systematic uncertainty are similar to the unmerged-jet category. The dominant experimental uncertainties are the background normalization and shape uncertainty, particularly in the 2-jet category, where the number of events is small, as well as the uncertainty in the signal cross section due to the hadronic W candidate selection. The dominant theoretical uncertainty comes from the uncertainty in the cross section when dividing the analysis into exclusive jet multiplicity bins.

H → ZZ → 2 2 '
This analysis seeks to identify Higgs boson decays to a pair of Z bosons, with one Z decaying to a pair of muons or electrons (Z → 2 , with = µ or e), and the second decaying to electrons,

CMS
, 2-jet µ e+ merged-jet category Figure 3: The final WW invariant mass distribution is shown for the 0+1-jet bin category for the muon channel only (left) and for the 2-jet bin category (right). Points represent the observed data, shaded graphs represent the background and dashed graphs represent five times the expectation for a m H = 800 GeV SM-like Higgs boson from ggF and VBF production, separately. muons or taus (Z → 2 ', with ' = e, µ or τ) [11,19,110]. This channel has extremely low background, and the presence of four leptons in the final state allows reconstruction and isolation requirements to be very loose, leading to a high selection efficiency. This channel is one of the most sensitive channels across the entire mass range.
For this analysis, we require triggers with two high-p T muons or electrons. The dilepton trigger p T thresholds for the leading and trailing leptons were 17 and 8 GeV, respectively. Events included in the analysis contain Z boson candidates formed from a pair of leptons of the same flavor and opposite charge. Decay muons or electrons are required to be isolated, and to originate from the primary vertex. Muons (electrons) are required to have p T > 5(7) GeV and |η| < 2.1(2.5), while taus are required to have a visible transverse momentum p T > 20 GeV and |η| < 2.3. We reconstruct the Z → ττ in the following decay modes: Z → τ h τ h , Z → τ e τ h , Z → τ µ τ h , and Z → τ e τ µ . Overlap with the 2 2 channel is avoided by excluding events with both taus decaying to electrons or muons.
For the 2 2 final state, the lepton pair with invariant mass closest to the nominal Z boson mass, denoted Z 1 , is identified and retained if 40 < m Z 1 < 120 GeV. The second Z boson, denoted Z 2 , is then constructed from the remaining leptons in the event, and is required to satisfy 12 < m Z 2 < 120 GeV. If more than one Z 2 candidate remains, the ambiguity is resolved by choosing the leptons of highest p T . Amongst the four candidate decay leptons, at least one should have p T > 20 GeV, and another should have p T > 10 GeV. This requirement ensures that selected events correspond to the high-efficiency plateau of the trigger.
For the 2 2τ final state, events are required to have one Z 1 → 2 candidate, with one lepton having p T > 20 GeV and the other p T > 10 GeV. The leptons from leptonic decays of the tau are required to have p T > 10 GeV. The invariant mass of the reconstructed Z 1 candidate is required to satisfy 60 < m < 120 GeV. The Z 2 candidate mass reconstructed from the visible tau decay products (visible mass), m ττ , must satisfy 30 < m ττ < 90 GeV for events with at least one hadronically decaying tau, and 20 < m ττ < 90 GeV for events with two leptonically decaying taus. Events with both taus decaying to muons or electrons are excluded in order to avoid overlap with the 2 2 channel. The selections on the reconstructed Z 1 mass for the 2 2τ final state are tighter than for the 2 2 channel, because the search range starts at m H = 200 GeV rather than m H = 145 GeV.
In order to further separate signal from background and to distinguish different signal production mechanisms, we use a matrix element likelihood approach, which relies on probabilities for an event to come either from signal or background using a set of observables, such as angular and mass information, which fully characterize the event topology in its center-of-mass frame [64].
The events are categorized according to their jet multiplicity, counting jets with p T > 30 GeV. In the 0-or 1-jet category the p T spectrum of the four-lepton system is exploited to distinguish between ggF and vector boson induced production modes, such as VBF and associated production with a vector boson (VH). The events selected with two or more jets potentially come from several sources: background (predominantly ZZ+2 jets), VBF signal, VH signal (with V→2 jets), and ggF signal gg → H+2 jets. In order to isolate the production mechanism, we use a matrix-element-based probability [64,111] for the kinematics of the two jets and the Higgs boson to come either from the VBF process or from the gg → H+2 jets process. This discriminant is equally efficient to separate VBF from either H+2 jets signal or ZZ+2 jets background, because both processes show very similar jet kinematics.
To allow estimation of the tt, Z+jets, and WZ+jets backgrounds, a Z 1 +X control region is defined, well separated from the expected signal region. In addition, a sample of Z 1 + events, with at least one reconstructed lepton in addition to a Z, is defined in order to estimate the lepton misidentification probability, which is the probability for non-prompt leptons and other particles, which are not leptons, to be reconstructed as leptons and to pass the isolation and identification requirements. The contamination of the sample from WZ events containing a genuine additional lepton is suppressed by requiring the energy imbalance measured in the transverse plane to be below 25 (20) GeV for the 2 2 (2 2τ) channel. The event rates measured in the background control region are extrapolated to the signal region. The systematic uncertainties associated with the reducible background estimate vary from 30% to 100% and are combined in quadrature with the statistical uncertainties.
The cross section for ZZ production at NLO is calculated with MCFM. The theoretical uncertainty in the cross section is evaluated as a function of m 2 2 , varying both the renormalization and factorization scales and the PDF set, following the PDF4LHC recommendations [74,[105][106][107][108].
The reconstructed invariant mass distribution for 2 2 , combining the 4µ, 2µ2e, and 4µ channels, is shown in Fig. 4 (left), compared with the expectation from SM background processes. Figure 4 (right) plots the reconstructed visible mass distribution for the 2 2τ selection, combining all 2 2τ final states.
The background shapes are taken from simulation, with rates normalized to the observed data. The measured Z boson and ZZ masses are underestimated due to the undetected neutrinos in tau decays. To compensate for this effect we correct the momenta of the visible tau decay products to constrain the Z boson mass to 91.2 GeV. The correction is applied on the selected events and only affects the final mass shape of the ZZ system. The observed mass distributions are well-matched to the SM background expectation.
In the 2 2 channel limits on the production of heavy Higgs bosons are extracted using the unbinned four lepton mass distribution and the correlation of the kinematic discriminant with the Higgs boson mass. In the case of the tagged 2-jet category a third dimension is introduced using the correlation of the aforementioned VBF discriminant with the four-lepton mass. In the untagged 0/1-jet category the transverse momentum of the four lepton system is used in place of the VBF discriminant. For the 2 2τ final state, limits are set using the m 2 2τ distribution.

H → ZZ → 2 2ν
This analysis seeks to identify Higgs boson decays to a pair of Z bosons, with one Z boson decaying to neutrinos and the other decaying to leptons. The analysis strategy is based on selection requirements in the (E miss T , m T ) phase space, with selections adjusted for different m H hypotheses [112]. Here, the transverse mass is determined between the transverse dilepton system, p T , and the E miss T . As in the previous Section, we use data collected with the trigger requiring two high-p T electrons or muons. Events are required to have a pair of well-identified, isolated leptons of same flavor (µ + µ − or e + e − ), each lepton with p T > 20 GeV, with an invariant mass within a 30 GeV window centered on the Z boson mass. The p T of the dilepton system is required to be greater than 55 GeV. The presence of large E miss T (70 GeV or more, depending on m H ) in the event is also required.
To suppress the Z+jets background, events are rejected if the angle in the transverse plane between the E miss T and the closest jet with p T > 30 GeV is smaller than 0.5 radians. Events where the lepton is mismeasured are rejected if E miss T > 60 GeV and ∆φ( , E miss T ) < 0.2. The top quark background is suppressed by applying a veto on events having a b tagged jet with p T > 30 GeV and |η| < 2.5. To further suppress this background, a veto is applied on events containing an additional, soft muon with p T > 3 GeV, typically produced in the leptonic decay of a b quark. To reduce the WZ background, in which both bosons decay leptonically, any event with a third lepton (µ or e) with p T > 10 GeV, and passing the identification and isolation requirements, is rejected.
The search is carried out in two mutually exclusive categories. The VBF category contains events with two or more jets in the forward region, with a |∆η jj | > 4 requirement between the two leading jets, and with the invariant mass of those two jets greater than 500 GeV. In addition, the two leptons forming the Z candidate are required to lie between these two jets in η, while no other selected jets with p T > 30 GeV are allowed in this region. The ggF category includes all events failing the VBF selection, and is subdivided into subsamples according to the presence or absence of reconstructed jets with p T > 30 GeV. The event categories are chosen in order to maximize the expected cross section limit. In the case of the VBF category, a constant E miss T > 70 GeV and no m T requirement are used, as no gain in sensitivity is obtained with a selection dependent on the Higgs boson mass hypothesis. In the case of ggF, we apply an m H -dependent lower limit on E miss T ranging from 80 to 100 GeV over the search range.
The background composition is expected to vary with the m H hypothesis. At low-m H , Z+jets and tt processes are the largest contributions, while at m H > 400 GeV the irreducible ZZ and WZ backgrounds dominate. The ZZ and WZ backgrounds are estimated from simulation and are normalized to their respective NLO cross sections. The Z+jets background is modeled from a control sample of events with a single photon produced in association with jets. This procedure yields a good model of the E miss T distribution in Z+jets events.
The uncertainty associated with the Z+jets background is affected by residual contamination of the γ+jets control sample from processes involving a photon and genuine E miss T . We do not explicitly subtract this contamination, but include a shape uncertainty in the final fit, which is allowed to vary this small residual background between 0 and 100% of this estimate.
Background processes that do not involve a Z boson (nonresonant background), include tt, single top quark production, W+jets and WW, and are estimated with a control sample of DF dilepton events (µ ∓ e ± ) that pass the full event selection. This method cannot distinguish between the nonresonant background and a possible contribution from H → WW → 2 2ν events, which are treated as part of the nonresonant background estimate. The nonresonant background in the µ + µ − or e + e − final states is estimated by applying a scale factor to the selected µ ± e ∓ events, obtained from the events in the sidebands of the Z boson peak (40 < m < 70 GeV and 110 < m < 200 GeV). The uncertainty associated with the estimate of this background is determined to be 25%. No significant excess of events is observed over the SM background expectation.
The m T and E miss T distributions used to extract the results are shown in Fig. 5 for all three event categories with the muon and electron channels combined.

H → ZZ → 2 2q
The H → ZZ → 2 2q channel has the largest branching fraction of all H → ZZ channels under consideration, but also a large background contribution from Z+jets production. Furthermore, the reconstruction of the hadronically decaying Z boson is more difficult than the fully leptonic final states. The dominant background is due to the DY process and can be reduced using b-tagging requirements on the selected jets, as will be described later in this section.
As in the other channels with a Z boson in the final state, we use data collected with the trigger requiring two high-p T muons or electrons. Reconstructed muons and electrons are required to have p T > 40 (20) GeV for the leading (second) lepton. Muons are required to have |η| < 2.4, and electrons |η| < 2.5. Jets are required to have p T > 30 GeV and |η| < 2.4. Each pair of oppositely charged leptons of the same flavor and each pair of jets are considered as Z boson candidates.
To increase the sensitivity to a possible signal, the main analysis is complemented with dedicated selections for the VBF signature. VBF events are identified requiring two additional jets with p T > 30 GeV, m jj > 500 GeV, and |∆η jj | > 3.5.
Analogously to the H → WW → νqq channel, the analysis is modified for very high Higgs boson masses (m H > 600 GeV) to account for the fact that the two jets originate from a Lorentzboosted Z boson and therefore they may be reconstructed as a single, merged jet (Sections 4 and 5.2.2). Information about the internal structure of this kind of jet [96] is used in order to gain some insight about the origin of the jet, distinguishing the DY background from jets produced from boosted Z bosons. To reduce the contamination from nonboosted Z + X backgrounds, the selection for the merged-jet topology requires the hadronically decaying Z boson to have p T > 100 GeV and |η| < 2.4, and the leptonically decaying Z boson to have p T > 200 GeV.
In order to exploit the different jet composition of signal and background, events are classified into three mutually exclusive categories according to the number of selected b tagged jets: 0 b tag, 1 b tag, and 2 b tag. This distinction is not done in the VBF oriented selection where the main discrimination is given by a multivariate discriminator for VBF topology based on angular and energy information of the two VBF tag jets.
Background contributions are reduced by requiring 71 < m jj < 111 GeV and 76 < m < 106 GeV in the selected events. The presence of Z bosons decaying to leptons and dijets makes this selection very efficient for signal, whereas the continuous background gets largely reduced. In case of the merged-jet analysis, the dijet requirement is applied on the merged-jet mass, after applying the pruning procedure.
An angular likelihood discriminant is used to separate signal-like from background-like events in each category [64,113]. In order to suppress the substantial expected tt background in the 2 b tag category, a discriminant is used, defined as the logarithm of the likelihood ratio of the hypothesis that the E miss T is equal to the value measured by the PF algorithm and the null hypothesis E miss T = 0 [84]. This discriminant provides a measure of whether the event contains genuine missing transverse energy. When an event contains multiple Z boson candidates passing the selection requirements, those with jets in the highest b tag category are retained for analysis. If multiple candidates are still present, the ones with m jj and m values closest to the Z boson mass are retained. In the case of the merged-jet category all the requirements on jets with respect to b tagging and the likelihood discriminant are applied to the two subjets reconstructed inside the merged jet. In case of events with a merged-jet Z boson candidate in addition to a dijet one, the candidate from the merged jet gets selected.
The dominant Z+jets background is estimated using simulated events properly corrected to reproduce the yield of the observed data in the control regions, defined as 60 < m jj < 71 GeV and 111 < m jj < 130 GeV. Simulated events are weighted to reproduce the p T spectrum of the jj system in these control regions. In case of the merged-jet category both the p T of the and J systems have weights applied. The normalization of this distribution is taken from observed data and used as an additional constraint for this background.
Other backgrounds are estimated using simulated events. These include diboson and top quark events. In the VBF selection, the background from top quark events is small. In the ggF analysis it is estimated from observed data using a control sample of eµ events, invoking lepton flavor symmetry that is satisfied by several background contributions but not in the signal events. The distributions of m ZZ in the signal region are shown in Fig. 6 for the three b tag multiplicities of the dijet category comparing observed data with background expectations. Good agreement is observed within the uncertainties. For the merged and VBF categories good agreement is also observed. The dominant systematic uncertainties are due to the b-tagging performance and the JES [90]. Further systematic effects are due to the uncertainty in the predicted signal and background shapes used in the analysis. The distributions of m ZZ in the signal region are used to extract the final signal and background yields.

Systematic uncertainties
Systematic uncertainties for the various final states come from common treatment of the signal model assumptions, reconstructed objects used in the analysis and a few common experimental effects.
Uncertainties on the cross section for the production of heavy Higgs bosons arise from uncertainties in the combined choice of the PDFs and α s , as well as in the renormalization and factorization scales [66], which are typically 6-7% and 7-12%, respectively, for the ggF production mechanism, and 1-2% and 2-5%, respectively, for production via VBF. Additionally, we add an uncertainty in the background coming from off-shell h(125) production, which we estimate with GG2ZZ (PHANTOM) for the ggF (VBF) case. We find that at the largest m H values, the size of the effect is approximately 3% of the total background. Uncertainties on the signal line-shape reweighting with interference varies for the ggF and VBF modes. For ggF, we follow the prescription in Ref. [66], which considers the NNLO contribution to the signal interfering with the gg → ZZ background process. For VBF, without a full prescription, we assign systematic uncertainties coming from renormalization and factorization scale variations in the PHANTOM generator.
Other common systematic effects come from the luminosity uncertainty, which is 2.2% (2.6%) for the 7 (8) TeV data. Uncertainties on the muon and electron reconstruction efficiencies, and JES and jet energy resolution (JER) are correlated among the various final states, where all these effects are subdominant. The lepton fake rate is largest for the H → ZZ → 2 2 channel, in which we consider fake leptons at relatively lower p T than in the other channels. A summary of the systematic uncertainties per channel is given in Table 2. Table 2: Sources of systematic uncertainties considered in each of the channels included in this analysis. Uncertainties are given in percent. Most uncertainties are affecting the normalisation of the observed data or simulated events, but some are uncertainties on the shape of kinematic distributions. Wherever ranges of uncertainties are given, they are either ranges in m H , jet multiplicity categories, or dependent on the production mode.

Statistical interpretation
The combination of the measurements in the different channels presented in this paper requires the simultaneous analysis of the data selected by all individual analyses, accounting for all statistical and systematic uncertainties, as well as their correlations. The statistical methodology used in this combination was developed by the ATLAS and CMS Collaborations in the context of the LHC Higgs Combination Group [12,114,115]. Upper limits on the model parameters are set for different m H hypotheses using a modified frequentist method referred to as the CL s method [116,117], where a likelihood ratio test statistic is used in which the nuisance parameters are profiled. In the likelihood ratio, the total number of observed events is compared to the signal and background predictions by means of a product of Poisson probabilities. The predictions are subject to the multiple uncertainties described in the previous section, and are handled by introducing nuisance parameters with probability density functions. The nuisance parameters modify parametrically the expectations for both signal and background processes. Furthermore, a signal strength modifier (µ) is used to scale the Higgs boson cross sections of all production mechanisms by the same factor with respect to their SM predictions while keeping the decay branching fractions unchanged.

SM-like Higgs boson search
The combined results obtained for a heavy Higgs boson with SM-like couplings for all the different contributing final states are displayed in Fig. 7. On the left, the observed 95% CL limit is shown for each final state. The expected combined 95% CL limit of the six channels is plotted as a dashed black line, while the yellow shaded region is the ±2σ uncertainty in the expected limit. On the right, the expected and observed limits are displayed for each of the individual channels as well as the combined result. Features in the combined observed limit can be traced to corresponding features in the limits of the individual channels. At lower masses below 400 GeV, there are oscillations in the observed limit due to the high resolution channel, H → ZZ → 4 , and the narrow width of the heavy Higgs boson in this mass range. An excess in the combined limit at around 280 GeV is related to a small excess in the channels H → ZZ → 4 and H → WW → 2 2ν. The small excess of observed events seen around 700 GeV in the H → WW → νJ merged-jet category is not supported by the other channels and is reduced to less than 0.5 σ in the combination. The combined upper limits at the 95% CL on the product of the cross section and branching fractions exclude a Higgs boson with SM-like couplings in the full search range 145 < m H < 1000 GeV.

EW singlet Higgs boson search
We interpret the search results in terms of a heavy Higgs boson in the EW singlet extension of the SM. The parameters of the model are C 2 , the heavy Higgs boson contribution to electroweak symmetry breaking, and B new , the contribution to the Higgs boson width of non-SM decays, and are defined in Section 3. Figure 8 shows the expected and observed upper limits at 95% CL on the singlet scalar cross section with respect to its expected cross section as a function of its mass.
The region above the curves shows the parameter space that is expected to be excluded at 95% CL. We show the exclusion region for various values of B new . We find a large region of C 2 versus mass parameters to be excluded for various values of B new . We also plot the µ h(125) = 1 ± 0.14 [78] indirect constraint C 2 < 0.28 at 95% CL for B new =0. The upper dashdotted line shows the cutoff of the allowed region for B new = 0.5 where the width of the heavy Higgs boson becomes larger than the SM width at that mass hypothesis.
In order to understand the constraints of these results in a model-independent approach, we further subdivide the results into categories. In Fig. 9   rations. At the top of Fig. 9 are the limits we obtain when we combine the ZZ (top left) and WW (top right) channels separately. Since the ZZ channels are more sensitive in the search for a Higgs boson with SM-like couplings, they better constrain the BSM case as well. The bottom of Fig. 9 shows the combined 95% CL for all final states but only the ggF or VBF production mechanism for the heavy Higgs boson. In the heavy Higgs boson with SM-like couplings scenario, we assume the ratio of the cross sections for various production mechanisms to be the same as in the SM case.

Summary
Combined results are presented from searches for a heavy Higgs boson in H → WW and H → ZZ decay channels, for Higgs boson mass hypotheses in the range 145 < m H < 1000 GeV. In the case of a Higgs boson decaying into a pair of W bosons, the fully leptonic (H → WW → ν ν) and semileptonic (H → WW → νqq) final states are considered in the analysis. For a Higgs boson decaying into two Z bosons, final states containing four charged leptons (H → ZZ → 2 2 ), two charged leptons and two quarks (H → ZZ → 2 2q), and two charged leptons and two neutrinos (H → ZZ → 2 2ν) are considered, where = e or µ and = e, µ, or τ.
The observed data are interpreted both in the context of a heavy Higgs boson with SM-like couplings and decays, as well as a search for a heavy, narrow resonance as an EW singlet partner of the SM Higgs boson at 125 GeV. No significant excess over the expected SM background has been observed and exclusion limits have been set. In the case of the search for a heavy Higgs boson with SM-like couplings and decays, we exclude the existence of such a heavy Higgs boson over the entire search range of 145 < m H < 1000 GeV. For the EW singlet partner of the SM Higgs, the parameters of the model are C 2 , the heavy Higgs boson contribution to EW symmetry breaking, and B new , the contribution to the Higgs boson width of non-SM decays. We find that a large part of the C 2 versus mass parameter space is excluded for various values of B new . Additionally, we present limits for the EW singlet model for different production mechanisms, ggF and VBF, and WW and ZZ decay modes separately.