1 Introduction

The Higgs field’s Yukawa couplings explain in a satisfactory way, and in agreement with data, the generation of the charged-fermion masses. However, the precise origin of tiny but non-zero neutrino masses, brought to light by neutrino oscillation experiments [1, 2], is not explained by the Standard Model (SM) of particle physics and remains a mystery today. Among the most compelling solutions to this mystery is the hypothesised existence of Majorana masses for neutrinos, which can be explored by high- and low-energy experiments [3,4,5,6].

Majorana neutrino masses are present in beyond-the-SM (BSM) theories that include new, heavy leptons or extended scalar sectors, collectively known as Seesaw models, or in theories with extended gauge sectors (Left-Right Symmetric Model, Grand Unified Theories) [7]. Even if these new particles have masses that are much higher than the energies reached at the Large Hadron Collider (LHC), and are therefore kinematically inaccessible, the Majorana nature of active neutrinos can still be probed indirectly using effective field theories, with the lowest-dimensional operator being the so-called dimension \(d=5\) Weinberg operator [8].

At energies below 100 \(\text {MeV}\), the Majorana nature of neutrinos can be probed by searching for neutrinoless double beta decays of certain isotopes, which are characterised by the appearance of two same-sign electrons (e) with an absence of neutrinos (\(\nu \)) in the final state [9]. Although the production of same-sign leptons (\(\ell \)) involving muons (\(\mu \)) or \(\tau \)-leptons is kinematically forbidden in such isotope decays, the same process can be studied at high energies when mediated at dimension \(d=5\), 7, or higher  [10,11,12], as shown in Fig. 1a. This is the subject of the present work, which searches for signs of heavy neutral leptons in the same-sign dimuon final state with the ATLAS detector. This search also connects to existing LHC searches for heavy Majorana neutrinos in resonant production [13,14,15,16], with the goal of extending the reach for heavy neutrino masses to beyond the \(\text {TeV}\)  scale [10], where the resonant production channels become kinematically inaccessible. A similar search was reported recently by the CMS Collaboration [17]. Searches for heavy Majorana neutrinos lighter than the Z boson were conducted at LEP [18].

Fig. 1
figure 1

Diagrammatic representation of (a) same-sign \(\mu ^{+}\mu ^{+}\) production in \(W^{+} W^{+}\) scattering mediated by a Majorana neutrino N in proton–proton collisions and (b) electroweak same-sign \(W^{+}W^{+}\) scattering, which is the main background in this search. The corresponding diagrams with negative-charged particles are also relevant to this work but not included here for simplicity

This search is conducted in the context of two benchmark models. The first is the Phenomenological Type-I Seesaw model [3, 19], where heavy Majorana neutrinos couple to SM particles through mass-mixing with light neutrinos, which is parameterised with a complex-valued matrix element. For simplicity, this work investigates only the phenomenology of the lightest heavy-neutrino mass eigenstate N by assuming that all other heavy eigenstates are decoupled. The benchmark scenario is considered without e and \(\tau \) flavour mixing, so the only non-zero entry in the mass-mixing matrix is the element \(\vert V_{\mu N}\vert ^2\). In this case the signal cross section \(\sigma (pp \rightarrow \mu ^\pm \mu ^\pm +X)\) can be written as a function of the cross section \(\sigma _0(pp \rightarrow \mu ^\pm \mu ^\pm +X)\) for the flavour-aligned scenario (in which \(|V_{\mu N}|=1\)):

$$\begin{aligned} \sigma (pp \rightarrow \mu ^\pm \mu ^\pm +X) \equiv \vert V_{\mu N}\vert ^4 \times \sigma _0(pp \rightarrow \mu ^\pm \mu ^\pm +X). \end{aligned}$$
(1)

The two free parameters of the model are \(\vert V_{\mu N}\vert ^2\) and the mass \(m_N\) of the heavy Majorana neutrino.

The second benchmark model is the \(d=5\) Weinberg operator [8], which extends the SM Lagrangian by:

$$\begin{aligned} \mathcal {L}_{5} = \ \frac{C_5^{\ell \ell '}}{\Lambda } \big [\Phi \!\cdot \! \overline{L}^c_{\ell }\big ] \big [L_{\ell '}\!\!\cdot \!\Phi \big ], \end{aligned}$$

where \(\Lambda \) is the scale at which the particles mediating the BSM signal become relevant degrees of freedom; \(C_5^{\ell \ell '}\) is a complex, flavour-dependent Wilson matrix coefficient; \(L_\ell ^T=(\nu _\ell ,\ell )\) is the left-handed lepton doublet; and \(\Phi \) is the SM Higgs doublet. Considering the Higgs vacuum expectation value \(\upsilon =\sqrt{2}\langle \Phi \rangle \approx 246~\text {GeV} \), it is possible to define the quantity

$$\begin{aligned} m_{\ell \ell '} = C_5^{\ell \ell '} \upsilon ^2/\Lambda , \end{aligned}$$
(2)

which for \(\ell \ell ' = ee\) reduces to the effective Majorana mass usually quoted by neutrinoless double beta decay experiments, which is a measure of the scale at which lepton-number symmetry is broken [9, 11]. For simplicity, it is assumed here that \(C_5^{\mu \mu }\) is the only non-zero Wilson coefficient of this matrix. The signal cross section \(\sigma (pp \rightarrow \mu ^\pm \mu ^\pm +X)\) can then be written as a function of the dimensionless effective cross section \(\sigma _5(pp \rightarrow \mu ^\pm \mu ^\pm +X)\) for \(\Lambda /|C_5^{\mu \mu }|=1~\text {TeV} \) as follows:

$$\begin{aligned} \sigma (pp \rightarrow \mu ^\pm \mu ^\pm +X) \equiv \frac{|C_5^{\mu \mu }|^2}{\Lambda ^2} \cdot \sigma _5(pp \rightarrow \mu ^\pm \mu ^\pm +X). \end{aligned}$$
(3)

The analysis targets final states with two same-sign muons, and also two particle jets with a large angular separation to capture the signature of vector-boson scattering. The main backgrounds are SM \(W^{\pm }W^{\pm }jj \) production, shown in Fig. 1b, and WZ production, where the bosons decay leptonically. They are estimated with simulated events and their modelling is constrained in dedicated signal-depleted control regions. A data-driven technique assisted by simulation is used to model backgrounds which do not originate from prompt muons. The presence of a signal is tested in a region designed to be signal-enriched using the benchmark models discussed above.

2 The ATLAS detector

The ATLAS detector [20] at the LHC is a multipurpose particle detector with a forward–backward symmetric cylindrical geometry with a nearly \(4\pi \) coverage in solid angle.Footnote 1 It consists of an inner tracking detector (ID) surrounded by a superconducting solenoid, electromagnetic (EM) and hadron calorimeters, and a muon spectrometer (MS).

The ID covers the pseudorapidity range \(|\eta | < 2.5\). The high-granularity silicon pixel detector covers the vertex region and typically provides four measurements per track, the first hit normally being in the insertable B-layer installed before Run 2 [21, 22]. It is surrounded by a silicon microstrip detector and a straw tube transition-radiation tracking detector. The calorimeter system covers the pseudorapidity range \(|\eta | < 4.9\). It consists of lead/liquid-argon (LAr) sampling calorimeters which provide EM energy measurements with high granularity up to \(|\eta | = 3.2\). A steel/scintillator-tile hadron calorimeter covers the central pseudorapidity range (\(|\eta | < 1.7\)). The endcap and forward regions are instrumented with LAr calorimeters for EM and hadronic energy measurements up to \(|\eta | = 4.9\). The MS surrounds the calorimeters and is based on three large air-core toroidal superconducting magnets with eight coils each. The field integral of the toroids ranges between 2.0 and 6.0 Tm across most of the detector. The MS includes a system of precision chambers for tracking and fast detectors for triggering. The ATLAS trigger and data acquisition system [23] contains a two-level trigger system in order to select events of interest. The first-level trigger is implemented in hardware and uses a subset of the detector information to accept events at a maximum rate of nearly 100 kHz. The second-level is a software-based trigger that reduces the accepted event rate to 1 kHz, on average, depending on the data-taking conditions [23]. An extensive software suite [24] is used in data simulation, in the reconstruction and analysis of real and simulated data, in detector operations, and in the trigger and data acquisition systems of the experiment.

3 Object reconstruction and event selection

This analysis uses a set of pp collision data events collected by the ATLAS detector between 2015 and 2018 at a centre-of-mass energy of \(\sqrt{s}=13\) \(\text {TeV}\). Only events for which all detector subsystems were operational are considered. The data set corresponds to an integrated luminosity of 140 fb\(^{-1}\)  [25, 26].

This analysis is based on events where the detector readout is triggered by the presence of at least one muon, referred to as ‘single-muon triggers’. Single-muon triggers require the presence of a single muon with a minimum transverse momentum (\(p_{\text {T}}\)) in the range 20–26 \(\text {GeV}\) depending on the data-taking period [27].

Events are required to have at least one collision vertex reconstructed from at least two ID tracks with \(p_{\text {T}} > 500~\text {MeV} \). For events with several collision vertices, the one with the largest sum of the squared transverse momenta of the associated tracks is taken as the primary vertex [28].

Three types of muon candidates are used to select events for the signal region, to estimate the background contribution from fake muons, and to veto additional muons in the event selection. These muon candidates have different selection criteria and are referred to as ‘signal muons’, ‘spurious muons’, and ‘baseline muons’, respectively. Baseline muon candidates are reconstructed by combining tracks in the ID with tracks in the MS [29] and are required to have \(|\eta |<2.5\) and \(p_{\text {T}} > 3~\text {GeV} \). Further selections on the longitudinal and transverse impact parameters are imposed. The transverse impact parameter \(d_0\), measured from the beamline and divided by its estimated uncertainty, must satisfy \({|d_{0} |/\sigma (d_{0})} < 15\). The longitudinal impact parameter \(z_0\), measured from the primary vertex, is required to satisfy \({|z_{0} \sin (\theta ) |<1.5}\) mm. Baseline muon candidates must satisfy the ‘Loose’ cut-based identification working point defined in Ref. [30], based on the numbers of hits in the different ID and MS subsystems, and on the significance of the charge-to-momentum ratio q/p. To select signal muons, these selection criteria are tightened to \(p_{\text {T}} > 10\) \(\text {GeV}\), \(|\eta |<2.5\), \({|d_{0} |/\sigma (d_{0})}<3\) and \({|z_{0} \sin (\theta ) |<0.5}\) mm, and used together with the ‘Medium’ cut-based identification working point. In addition, an isolation is required where the scalar sum of the \(p_{\text {T}} \) of tracks inside a variable-size cone around the muon (excluding its own track) plus 40% of the transverse energy of neutral particle-flow objects in a cone of size \(\Delta R = 0.2\) around the muon, must be less than 4.5% of the muon \(p_{\text {T}} \) [30]. To select spurious muons, the above isolation requirement is loosened to be less than 16% of the muon \(p_{\text {T}} \) and \({|d_{0} |/\sigma (d_{0})}<10\). In addition, they must fail the tighter signal-muon isolation or transverse impact parameter requirements.

Electron candidates are reconstructed from an isolated EM calorimeter energy deposit which is matched to a track in the ID [31] and they are required to have \(p_{\text {T}} > 4.5~\text {GeV} \). The pseudorapidity of the calorimeter energy cluster, \(\eta _{\text {cluster}}\), must satisfy \(|\eta _{\text {cluster}}|<2.47\). Electron candidates must pass a ‘Loose’ likelihood-based identification working point [31] that employs calorimeter and tracking information to discriminate between electrons and jets, and their tracks must have impact parameters satisfying \({|d_{0} |/\sigma (d_{0})} < 5\) and \({|z_{0} \sin (\theta ) |<0.5}\) mm. In this analysis, electron candidates are used only to veto events.

Jets are reconstructed by using the anti-\(k_{t}\) algorithm [32, 33] with a radius parameter of \(R=0.4\) after applying a particle-flow algorithm [34] to tracking information and topological clusters of calorimeter-cell energies. Jets are required to have \(p_{\text {T}} > 25~\text {GeV} \) and \(|\eta | < 4.5\), and are calibrated as described in Ref. [35]. The jet-vertex tagger (JVT) discriminant [36] is used to identify jets originating from the hard-scatter interaction through the use of tracking and vertexing information. Jets with \(p_{\text {T}} < 60~\text {GeV} \) and \(|\eta | < 2.5\) are required to satisfy the requirement \(\text {JVT} > 0.50\), corresponding to a selection efficiency of about 96% for hard-scatter jets [37]. If either of the two leading jets has \(|\eta |{\,>\,}2.5\) and \(p_{\text {T}} < 60~\text {GeV} \), it must satisfy a threshold for the forward jet-vertex tagger (fJVT) discriminant [38, 39] that has an efficiency of about 93% for hard-scatter jets.

In order to suppress contributions from background processes that involve top quarks or leptonic b-hadron decays, the DL1r classification algorithm based on recurrent neural networks [40, 41] is used to identify jets containing b-hadrons (referred to as ‘\(b\)-jets’). This algorithm identifies \(b\)-jets amongst a background of light-flavour- and charm-quark-initiated jets using information about the impact parameters of tracks associated with the jet and the topological properties of displaced vertices reconstructed within the jet. This analysis uses a \(b\)-jet efficiency working point of 85% as measured for jets with \(p_{\text {T}} > 20~\text {GeV} \) and \(|\eta | < 2.5\) in simulated \(t\bar{t}\) events [42]. The tagging algorithm at that working point gives an expected rejection factor (defined as the inverse of the efficiency) of about 40 for light-flavour jets, and about 2.9 for jets originating from charm quarks [41, 43].

To resolve the potential ambiguities due to a single detector response being assigned to two objects by the reconstruction algorithm, a sequential overlap-removal procedure is applied. First, any electron found to share a track with another electron with higher \(p_{\text {T}}\) is removed. Second, electrons sharing their track with a muon candidate are removed. Third, labelling the rapidity of an object by y,Footnote 2 any jet within \(\Delta R_y = \sqrt{(\Delta y)^2+(\Delta \phi )^2} = 0.2\) of an electron is removed, unless it is a b-jet and the electron \(p_{\text {T}}\) is below 100 \(\text {GeV}\), in which case the electron is removed. Fourth, any electrons within \(\Delta R_y = 0.4\) of a remaining jet are removed. Fifth, any jet that is within \(\Delta R_y = 0.2\) of a muon and has less than three associated tracks is removed, unless it is a b-jet, in which case the muon is removed. Finally, remaining muons are removed if their track is within \(\Delta R_y = 0.4\) of a remaining jet. Resolving the overlaps between \(e/\mu \) and jets in two steps based on \(\Delta R_y\) values of 0.2 and 0.4 allows the reconstruction efficiency to be maximised for b-jets including semileptonic c- and b-hadron decays.

The missing transverse momentum, with magnitude \(E_{\text {T}}^{\text {miss}} \), is defined as the negative vectorial sum of the transverse momenta of all selected and calibrated electrons, muons and jets, and an additional term to account for soft energy in the event that is not associated with any of the selected objects. This soft term is calculated from unassociated ID tracks matched to the primary vertex. While it suffers from an inherent underestimation, which is covered by dedicated uncertainties, it is more resilient to contamination from multiple interactions in the same or neighbouring bunch crossings (pile-up) compared to the calorimeter-based soft term [44, 45]. The object-based \(E_{\text {T}}^{\text {miss}}\) significance (\(\mathcal {S}\)) is computed [46] by including the expected resolutions for all the objects used in the \(E_{\text {T}}^{\text {miss}}\) calculation and is used to reject background events with genuine \(E_{\text {T}}^{\text {miss}}\) originating from undetected particles such as neutrinos.

At preselection level, the events are required to have at least two same-sign signal muons with \(p_{\text {T}} > {27}\,{\textrm{GeV}}\) and one of them has to match the muon that fired the trigger. At least two jets are required and one of the jets has to have a transverse momentum larger than 30 GeV to reduce the contamination from pile-up jets. Events with b-jets are vetoed. In order to preferentially capture the signature of \(W^{\pm }W^{\pm }\) scattering in pp collisions, the invariant mass of the two jets with the highest \(p_{\text {T}}\), labelled as \(m_{jj}\), has to be larger than 300 \(\text {GeV}\) and their rapidity separation must satisfy \(|\Delta y_{jj}| > 4\). These preselection cuts are applied to every event which is used further in the analysis.

4 Event simulation

Monte Carlo (MC) samples are used to model the expected signal and background yields, while backgrounds from misidentified objects are estimated using data-driven techniques described in Sect. 5. All MC samples are normalised to the best theory predictions available. The NNPDF3.0nlo [47] PDF set was used in all matrix element calculations if not stated otherwise. The generated event samples were processed through the full ATLAS detector simulation [48] based on Geant4  [49]. Additional simulated \(pp\) collisions generated using Pythia8.186 [50] with the A3 set of tuned parameters (tune) [51] and the MSTW2008lo parton distribution function (PDF) set [52] were overlaid to model the effects of pile-up. The distribution of the number of additional \(pp\) interactions in the MC samples is reweighted to match the one observed in data. All simulated samples are processed through the same reconstruction algorithms and analysis chain as the data.

The Phenomenological Type-I Seesaw model signal samples were simulated following the prescription given in Ref. [10]. The samples were simulated using the MadGraph5_aMC@NLO2.7.2 [53] generator at next-to-leading order (NLO) in QCD after importing the default variant of the HeavyN [54, 55] FeynRules UFO libraries [56, 57]. All HeavyN events were interfaced to Pythia8.243 [58]. Twenty mass points were generated, spanning the range \(m_{N} = 50\) \(\text {GeV}\)  to 20 \(\text {TeV}\). Since the mixing parameter \(V_{\mu N}\) only impacts the total cross section, as \(\sigma \sim |V_{\mu N}|^{4}\), these mass points were generated with a mixing parameter of unity and are rescaled to other values by using Eq. (1).

The signal sample for the Weinberg operator was simulated by using the prescription given in Ref. [11]. For the hard parton-level scattering processes, MadGraph5_aMC@NLO 2.9.3 [53] was used at NLO in QCD after importing the default variant of the SMWeinberg [11] FeynRules UFO libraries [56, 57]. The NNPDF3.1luxQED [47] PDF was used. All SMWeinberg events were interfaced to Pythia8.245 [58]. The sample was generated at an energy scale \(\Lambda =200\) \(\text {TeV}\) and with the Wilson coefficient value \(C_5^{\mu \mu }=1\). Both \(\Lambda \) and \(C_5^{\mu \mu }\) only impact the total cross section, as \(\sigma \sim (C_5^{\mu \mu }/\Lambda )^2\), so a global rescaling of a single signal sample according to Eq. (3) is used to vary their values.

For all samples of simulated signal events, the parton shower, hadronisation, and underlying-event modelling by Pythiaused the A14 tune [59] and the NNPDF2.3lo [47] PDF set, and the EvtGen1.7.0 program [60] was used to model the decays of bottom and charm hadrons.

Electroweak (EW) \(W^\pm W^\pm \) production and decay in association with two jets was simulated with diagrams including exactly six orders of the EW coupling [61]. The simulation of the strong production processes, and the interference between EW and strong contributions, includes diagrams with exactly four and five EW vertices [62], respectively. All SM \(W^{\pm }W^{\pm }jj \) contributions were simulated with MadGraph5_aMC@NLO2.6.7 [53], which was interfaced to the Pythia8.244 [58] parton shower model. The matching of the matrix element simulation to parton showers used the MC@NLO method [63] based on FKS subtraction [64]. These samples used the A14 set of tunable parameters [59]. More details of the generator settings used to obtain a better description of deep-inelastic scattering data [65] are given in Ref. [61].

Samples of diboson final states (VV, \(V=W/Z\)) were simulated with the Sherpa2.2.1 or 2.2.2 [66] generator depending on the process. Fully leptonic final states and semileptonic final states, where one boson decays leptonically and the other hadronically, were generated using matrix elements with NLO accuracy in QCD for up to one additional parton and with LO accuracy for up to three additional parton emissions. Samples for the loop-induced processes \(gg \rightarrow VV\) were generated using LO matrix elements for up to one additional parton emission. The production of triboson (VVV) events was simulated with the Sherpa2.2.2 generator, accurate to NLO in QCD for the inclusive process and to LO for up to two additional parton emissions, using factorised gauge-boson decays. The matrix element calculations for individual VV and VVV processes were matched and merged with the Sherpaparton shower based on Catani–Seymour dipole factorisation [67, 68] using the MEPS@NLO prescription [69,70,71,72]. The virtual QCD corrections for matrix elements with NLO accuracy were provided by the OpenLoops1 library [73,74,75].

Additional samples are used to simulate minor backgrounds with prompt muons. The production of \(t\bar{t}V\), \(t\bar{t}WW\), \(tZq\) and \(tWZ\) events was modelled using the MadGraph5_aMC@NLO2.3.3 generator at NLO, interfaced to Pythia [58]. The Sherpa2.2.11 generator was used to simulate \(V\gamma \) final states. For studies of background which does not originate from prompt muons, the production of \(t\bar{t}\) and single-top-quark events was simulated using the Powheg Boxv2 [76,77,78,79] generator at NLO in QCD, interfaced to Pythia8.230 [58], and the production of \(V+\)jets events was simulated with the Sherpa2.2.11 generator.

Table 1 Summary of the main kinematic requirements for the different selection regions. The first four requirements belong to the preselection stage and are described in Section 3

5 Analysis strategy

The signal considered in this search is characterised by two same-sign muons with large transverse momentum and two jets with a large rapidity separation, which is already exploited in the preselection requirements. The signal region (SR) was optimised for the Phenomenological Type-I Seesaw model with Majorana neutrino masses above 1 \(\text {TeV}\), since this is the kinematic region where this search is expected to bring unique sensitivity. The best discrimination between signal and background is given by the distribution of the transverse momentum of the second-highest-\(p_{\text {T}}\) muon (\(p_\text {T}^{\mu _2}\)). The distribution has a maximum above 200 \(\text {GeV}\) for \(m_N \ge 500\) \(\text {GeV}\), with a long tail towards higher momenta [10], while all SM backgrounds have a steeply falling \(p_\text {T}^{\mu _2}\) distribution. The main background contributions arise from SM \(W^{\pm }W^{\pm }jj \) and WZ production. The SR is defined by requiring the \(E_{\text {T}}^{\text {miss}}\) to have a small significance value \(\mathcal {S} < 4.5\), and events with a third baseline muon or electron are vetoed. This suppresses background from SM \(W^{\pm }W^{\pm }jj \) events, which have two neutrinos in the final state (cf. Fig. 1b), and WZ events, which have a third lepton. In contrast, these two requirements are inverted to define signal-depleted control regions (CR) where the data are used to constrain the normalisation of these simulated background processes. The event selection for the SR and the two CRs is summarised in Table 1.

A same-sign \(W^{\pm }W^{\pm }jj \) control region (ssWW-CR) is selected with \(\mathcal {S} > 5.8\). In order to suppress possible contamination from signal events, \(p_\text {T}^{\mu _2}\) is restricted to be below 120 \(\text {GeV}\). This event selection also ensures that background from misidentified jets is adequately suppressed. The ssWW-CR has a SM \(W^{\pm }W^{\pm }jj \) purity of 70% inclusively (64% for \(80<p_\text {T}^{\mu _2} <120\) GeV), with the main contamination coming from WZ events, whose modelling is, however, constrained in another dedicated control region. Events which fulfil the ssWW-CR requirement but with a \(E_{\text {T}}^{\text {miss}}\) significance \(4.5< \mathcal {S} < 5.8\) or with \(p_\text {T}^{\mu _2} >120\) \(\text {GeV}\) could have a non-negligible signal contribution. These two regions are not included in the profile likelihood fit, but are used to validate the modelling of SM \(W^{\pm }W^{\pm }jj \) and WZ events. The final background predictions and the collision data agree well in both validation regions.

The second-largest background in the SR is WZ production where both bosons decay leptonically and the Z-boson-decay lepton opposite in sign to the W boson falls outside the detector acceptance (mainly due to the muon spectrometer’s rapidity coverage). To select a control region enriched in WZ production (WZ-CR), the SR is modified such that events are required to have three signal muons, with one opposite-sign (OS) muon pair to be compatible with a Z boson decay. Events with an additional baseline muon are vetoed to reduce the contribution from ZZ production. Furthermore, a large three-lepton invariant mass (\(m_{\ell \ell \ell } > 100\) \(\text {GeV}\)) is required in order to reduce the fake-lepton contribution and to make the WZ-CR orthogonal to the regions where fake leptons from photons are studied, as explained later in this section. The WZ-CR has a WZ purity of 88% inclusively and 83% for \(p_\text {T}^{\mu _2} > 120\) GeV.

Fig. 2
figure 2

Comparison of the \(p_\text {T}^{\mu _2}\) distributions in data and the post-fit prediction under the background-only hypothesis in the different analysis regions: (a) SR, (b) ssWW-CR, and (c) WZ-CR. The shaded area surrounding the background expectation represents the total uncertainty from the post-fit computation. The ratio of the observed yields to the prediction of the background model post-fit is shown by the points in the bottom plot and the change in the prediction relative to the pre-fit yields is shown by the blue line. The ‘Other’ contribution consists mainly of ZZ and \(tZq\) events. Two benchmark signal samples are overlaid on top of the total background prediction in (a), and were obtained by using their theoretical cross sections. The last bin in (a) and (c) includes the overflow bin

Non-prompt muons, defined as originating from bottom- or charm-hadron, kaon or pion decays, constitute a non-negligible source of background in the analysis. Most of these muons arise from W+jets events and \(t\bar{t}\) events where one W boson (originating from a top-quark decay in the \(t\bar{t}\) case) decays leptonically. This background contribution is evaluated using the so-called fake-factor method [80], which is based on collision data. Fake factors are derived as a function of muon \(p_{\text {T}}\) in a dedicated region enriched in non-prompt muons which selects events that include a jet recoiling almost back-to-back against a fake muon candidate. In particular, the presence of W+jets events in this sample leads to a non-negligible contamination from prompt muons. This contamination is reduced by requiring the transverse mass formed by the non-prompt muon candidate’s \(p_{\text {T}}\) and the missing transverse momentum to be below 50 \(\text {GeV}\). The residual prompt-muon contribution, estimated from MC predictions, is subtracted from the data, and the fake factors are calculated as the ratio of non-prompt ‘signal’ to ‘spurious’ muon events. The background from non-prompt muons is then estimated by multiplying the fake factors by the data yields observed in the so-called ‘spurious-SR’, which has the exact same requirements as the SR except that the object definition of one of the muons is changed from ‘signal’ to ‘spurious’. Non-prompt contributions originating from either the leading or subleading muon candidates are considered. The contamination from prompt muons in the spurious-SR, originating mostly from same-sign WW and WZ events, is subtracted using MC predictions. The final data-driven estimate gives a low non-prompt-muon event yield in the different analysis regions, and hence this contribution to the \(p_\text {T}^{\mu _2}\) distribution is smoothed by fitting it with an exponential function. Alternative functions are used to estimate the systematic uncertainty due to the functional form choice.

Background from \(t\bar{t}V\), \(t\bar{t}WW\), \(tWZ\), \(tZq\), ZZ and \(V\gamma \) events is estimated from MC simulation and denoted as ‘Other’ backgrounds in the remainder of the paper, in which ZZ and \(tZq\) constitute the dominant contributions.

Photons originating from final-state radiation splitting into muons (\(\gamma ^*\rightarrow \mu \mu \)) were studied in a region selecting leptonic Z boson decays produced in association with an extra muon (which is the non-prompt candidate from final-state radiation) with \(m_{\ell \ell \ell }<100\) \(\text {GeV}\) to suppress the events with initial-state radiation. Their contribution to the SR was found to be negligible. Muons with the wrong reconstructed charge were studied with MC simulation and were also found to be negligible. The probability of misreconstructing two inelastic collisions, each contributing one of a pair of same-sign W bosons, as a single event was evaluated. A Gaussian function with a standard deviation as measured in Ref. [81] was used to model the longitudinal distribution of inelastic collisions and to compute the conditional probability. Less than 0.1 events from this source are expected to pass the analysis selection. Finally, single pp interactions where more than one set of partons from each proton interact (double parton scattering) to produce two W bosons with the same sign were considered. Although produced at a rate several times larger than that of coincidental W events, less than one such event is expected to pass the analysis selction, so it is also safe to ignore this process.

6 Results

This search is conducted using the binned \(p_\text {T}^{\mu _2}\) distributions. A combined profile likelihood fit is used in the SR and all CRs to obtain the final background model, as shown in Fig. 2, and make further statistical inferences regarding the presence of a signal. The likelihood function depends on the free-floating signal normalisation, the free-floating SM \(W^{\pm }W^{\pm }jj\) and WZ normalisations, and a set of nuisance parameters encoding the effect of systematic uncertainties. The normalisation corrections obtained by the fit (defined as the ratio between post-fit and pre-fit predictions) are \(1.17\pm 0.23\) for SM \(W^{\pm }W^{\pm }jj\) and \(0.92\pm 0.16\) for WZ backgrounds. The nuisance parameters, implemented as Gaussian constraints, allow variations of the expectations for signal and background according to the corresponding systematic uncertainties, whilst penalising the likelihood function for any deviation from the nominal expectations according to the specified constraints. The fit also reduces the impact of systematic uncertainties on the search sensitivity by exploiting the background-dominated CRs. The fit and the subsequent statistical analysis are performed for each signal hypothesis.

Uncertainties affecting the measurement which originate from statistical sources are considered together with systematic uncertainties related to the detector calibration and physics modelling. Uncertainty sources related to muon reconstruction and energy calibration, jet energy calibration, b-jet identification, rejection of pile-up jets, and missing transverse momentum reconstruction were studied. The uncertainty in the non-prompt muon background is evaluated by propagating both the systematic and statistical uncertainties in the fake factors and in the spurious-SR non-prompt muon yields. The systematic uncertainty is dominated by the modelling of the contamination from prompt muons in the region used to extract the fake factors. The statistical uncertainty, which dominates the uncertainty in the non-prompt background, is driven by the limited size of the spurious-SR data set. Systematic uncertainties in the physics modelling of SM same-sign WW and WZ events are estimated by varying the factorisation (\(\mu _\textrm{f}\)) and renormalisation (\(\mu _\textrm{r}\)) scales, as well as the strong coupling constant \(\alpha _{\text {s}}\) and the PDF choice. Factorisation and renormalisation scale variations are used to account for the uncertainty due to truncated higher-order corrections. The effects of a seven-point pairwise variation of \(\mu _\textrm{f}\) and \(\mu _\textrm{r}\), where each scale is either halved or doubled (the off-diagonal variations by [0.5, 2] and [2, 0.5] are not included), are computed and then combined into a single uncertainty by forming an envelope of the effects of these seven variations. The total PDF uncertainty is obtained from the standard deviation of NNPDF3.0nlo variations and the envelope of alternative PDFs (CT14 and MMHT2014), while the effect of the \(\alpha _{\text {s}}\) uncertainty is obtained by varying the nominal \(\alpha _{\text {s}} =0.118\) within its uncertainty of \(\pm 0.001\). Alternative SM same-sign WW (WZ) samples generated with Sherpa2.2.12 and MadGraph5_aMC@NLO interfaced to Herwig7.2 [82] (MadGraph5_aMC@NLO 2.2.2 interfaced to Pythia8.186) are used to evaluate the residual generator dependencies. A similar scheme is applied to derive uncertainties from the \(\mu _\textrm{f}/ \mu _\textrm{r} \) scale variations and from the PDF/\(\alpha _{\text {s}}\) choice for the HeavyN and SMWeinberg signal samples, while other smaller backgrounds are assigned a conservative \(\pm 50\%\) uncertainty correlated across analysis bins. The largest systematic uncertainty in the total background estimate is the uncertainty in the non-prompt-muon background estimation (8%). None of the considered systematic uncertainties were pulled appreciably by the fit or found to have an impact on the analysis sensitivity, which is driven by the statistical uncertainty of the data in the tail of the \(p_\text {T}^{\mu _2}\) distribution. The expected limits change at the per-mille level if systematic uncertainties are not considered.

Fig. 3
figure 3

Observed and expected 95% CL upper limits on the heavy Majorana neutrino mixing element \(\vert V_{\mu N}\vert ^2\) as a function of \(m_N\) in the Phenomenological Type-I Seesaw model. The one and two standard deviation bands of the expected limit are indicated in green and yellow, respectively

In the absence of any significant excess above the background expectation, upper limits on \(\vert V_{\mu N}\vert ^2\) and \(|C^{\mu \mu }_5| / \Lambda \) are derived with the \(\text {CL}_{\text {s}}\) method [83, 84]. The asymptotic approximation [85] is used to derive the upper limits, which agree very well with those estimated via background-only pseudo-experiments. The observed and expected 95% confidence level (CL) upper limits on the heavy Majorana neutrino mixing element \(\vert V_{\mu N}\vert ^2\) are shown in Fig. 3 as a function of \(m_N\). The form of the exclusion curve follows what is expected from the bare cross section \(\sigma _0\) as defined in Eq. (1), which rises moderately with increasing \(m_N\), reaching a maximum around 700 \(\text {GeV}\) for 13 \(\text {TeV}\) pp collisions before decreasing again due to a kinematic suppression in the \(W^{\pm }W^{\pm }\) matrix element [10]. The strongest exclusion is achieved at \(m_N=500\) \(\text {GeV}\), with an observed (expected) upper limit of 0.10 (0.08) on \(\vert V_{\mu N}\vert ^2\). For a heavy neutrino mass of 20 \(\text {TeV}\) the exclusion curve reaches the mixing element’s upper bound of unity. Overall, the expected limits constitute a sensitivity improvement of 15–20% compared to the results in Ref. [17]. Over the entire mass range the observed limit is around one standard deviation weaker than the expected limit, which is due to the slight excess of data in the last two bins of the \(p_\text {T}^{\mu _2}\) distribution. The observed (expected) 95% CL lower limit on the \(d=5\) Weinberg operator \(\Lambda / |C_5^{\mu \mu }|\) is 3.635 (4.608) \(\text {TeV}\), with a probability of \(p_0= 9.3\%\) that the observed data are compatible with the background-only hypothesis. Using Eq. (2), this translates into an upper limit of 16.7 (13.1) \(\text {GeV}\) on the effective \(\mu \mu \) Majorana neutrino mass.

7 Conclusion

A new search for heavy Majorana neutrinos is presented, targeting the same-sign WW production mode. Since in this process the heavy neutrinos are produced in the t-channel, this analysis extends the reach to higher neutrino masses and complements previous searches in resonant production channels by adding valuable sensitivity for neutrino masses of a few hundred \(\text {GeV}\). The search is performed using 140 fb\(^{-1}\) of pp collision data at \(\sqrt{s} = 13~\text {TeV} \) collected by the ATLAS detector at the LHC. To preferentially capture the signature of same-sign WW scattering in pp collisions, events are required to have two jets with a large rapidity separation and two muons with the same electric charge. Since signal events typically have a significantly higher muon transverse momentum than those arising from SM backgrounds, the shape of the transverse momentum distribution of the subleading muon is used to discriminate between the signal and background. No significant excess was observed above the expected backgrounds, which were modelled in dedicated control regions. Upper limits on the muon-neutrino–heavy-neutrino mass-mixing matrix element \(\vert V_{\mu N}\vert ^2\) at 95% confidence level are computed for signals with Majorana neutrino masses \(m_N\) between 50 \(\text {GeV}\) and 20 \(\text {TeV}\) based on a Phenomenological Type-I Seesaw model. The strongest exclusion was achieved at \(m_N=500\) \(\text {GeV}\), with an observed (expected) limit of 0.10 (0.08) on \(\vert V_{\mu N}\vert ^2\). Constraints on \(\vert V_{\mu N}\vert ^2\) are extended from previous results at \(m_N\) values of the order of 1 \(\text {TeV}\) in the resonant production channels to \(m_N\) values of up to 20 \(\text {TeV}\) in this analysis. In addition, limits are set on the \(d=5\) Weinberg operator, which translate to an observed (expected) upper limit of 16.7 (13.1) \(\text {GeV}\) on the effective \(\mu \mu \) Majorana neutrino mass, which cannot be probed in nuclear decays.