1 Introduction

While the Large Hadron Collider (LHC) pursues its quest for new physics, its Run 2 at a centre-of-momentum energy of 13 TeV being on-going, no direct evidence for physics beyond the Standard Model (SM) has been observed so far. The conceptual problems and limitations of the Standard Model therefore remain unsolved. Among the various theoretical frameworks tackling those issues, Supersymmetry still remains one of the most popular and appealing options. The absence of experimental evidence for any of the superpartners of the Standard Model degrees of freedom, and in particular of the strongly-interacting squarks and gluinos [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15], however imposes strong constraints on how Supersymmetry could be viably realised. For instance, the superpartners have to be too heavy to be produced at current LHC centre-of-mass energy and luminosity, or the particle spectrum has to be highly degenerate [16,17,18]. An alternative to these two solutions would be to abandon the idea of a minimal supersymmetric realisation which most current searches are based upon.

The ATLAS and CMS collaborations have obtained their limits under the assumption of a rather simplified realisation of the Minimal Supersymmetric Standard Model (MSSM). Those studies generally consider that a limited number of new physics states are light and can thus be produced at the LHC, and that those light particles therefore undergo a single well-defined decay mode. While studies of such simplified situations are reasonable and important starting points, the structure of the MSSM could be more general and complex and yield decay patterns that are not addressed by current searches.

In the present work, we address this last point and consider a realisation of the MSSM where inter-generational mixings in the squark sector are allowed. In this case, a physical squark eigenstate contains several components of well-defined flavour, which opens the door to multiple decay modes. For example, if the lightest squark is dominantly of top flavour but additionally contains a sizeable charm-flavour component, decays involving a top quark and a charm quark may both have a significant branching fraction. As a consequence, not all signatures stemming from squark pair production and decay are captured by the existing experimental searches, and current direct search limits may be weakened [19, 20].

In the MSSM realisations usually under consideration, squark inter-generational mixings are suppressed by imposing the minimal flavour violation (MFV) paradigm in which all flavour-violating interactions originate from the diagonalisation of the fermion sector and the corresponding CKM and PMNS matrices. Departing from an MFV squark sector, additional flavour-mixing terms may be present in the Lagrangian of the theory. By virtue of kaon data constraints, any mixing involving squarks of the first generation has to be extremely small [21], although second and third generation squark mixings are still largely allowed by current data [22]. We therefore focus on mixings solely involving the charm and top flavours of right-handed squarks. Such mixings can for instance originate from Grand Unification at a high scale [23, 24], and their TeV scale implications on squark production and decay have received considerable attention in the past [25,26,27,28,29,30,31,32,33,34,35,36,37]. In this work, we assess how they are constrained by the most recent LHC results and show, by proposing a novel class of search strategies, how future LHC searches could be tailored better to constrain non-minimally flavour-violating (NMFV) Supersymmetry.

The rest of this paper is organised as follows. In Sect. 2, we introduce the simplified setup that we have adopted to include inter-generational squark mixings in the MSSM. We then reinterpret the results of existing LHC direct searches for squarks when potentially altered squark decay modes are allowed. In Sect. 3, we present an analysis strategy targeting the specific channel where, after squark pair-production, one squark decays into a top quark and the second squark into a charm quark. Projections for sensitivity of the coming LHC runs to the model are presented and discussed in Sect. 4, where we emphasise the potential role of charm tagging. Our conclusions are given in Sect. 5.

2 Model setup and existing LHC limits

In this section, we describe the simplified model featuring top-charm mixing in the squark sector which we base our analysis on. We discuss general features of the search strategy for targeting such a model, and reinterpret recent squark searches to assess the current LHC coverage of the model parameter space.

2.1 A simplified model for squark flavour violation

In order to assess the LHC sensitivity to supersymmetric models featuring non-minimal flavour violation in the squark sector, we consider a simplified model embedding two active flavours of squarks, a right-handed top squark \(\tilde{t}_R\) and a right-handed charm squark \(\tilde{c}_R\). These two states mix into two physical eigenstates \(\tilde{u}_1\) and \(\tilde{u}_2\) whose flavour structure is dictated by the \(\theta _{tc}\) mixing angle,

$$\begin{aligned} \begin{pmatrix} \tilde{u}_1 \\ \tilde{u}_2 \end{pmatrix} = \begin{pmatrix} ~~\cos \theta _{tc} &{} ~\sin \theta _{tc} \\ -\sin \theta _{tc} &{} ~\cos \theta _{tc} \end{pmatrix} \begin{pmatrix} \tilde{c}_R \\ \tilde{t}_R \end{pmatrix}, \end{aligned}$$
(1)

where by convention \(\tilde{u}_1\) is the lighter of the two mass eigenstates. Our simplified model additionally includes one neutralino \({\tilde{\chi }}^0_1\), that we take bino-like. Such an assumption does not have a significant impact on our phenomenological results. The setup of our interest is thus based on four parameters: the masses \(m_{\tilde{u}_1}\) and \(m_{\tilde{u}_2}\) of the two physical squarks together with the flavour mixing angle \(\theta _{tc}\), and the neutralino mass \(m_{\tilde{\chi }^0_1}\). For the sensitivity studies in the \((m_{\tilde{u}_1}, \theta _{tc})\) plane, the neutralino mass will be fixed to \(m_{\tilde{\chi }^0_1}=50\,{\mathrm {\ GeV}}\). Although a more complicated flavour structure involving left-handed squarks could be possible as well, this last setup implies the need to handle more complicated constraints from B-physics in order to build phenomenologically viable scenarios. Left-handed squarks are thus assumed heavier and decoupled, like any other superpartner.

Our simplified model therefore exhibits two competing squark decay modes (if kinematically allowed),

$$\begin{aligned} \tilde{u}_i \rightarrow t {\tilde{\chi }}^0_1, \quad \tilde{u}_i \rightarrow c {\tilde{\chi }}^0_1 \quad \text {with}\ \ i=1,2, \end{aligned}$$
(2)

which yield three classes of LHC signatures originating from the production of a pair of \(\tilde{u}_i\) squarks. Typical LHC search strategies have been designed on the basis of the MFV paradigm and thus only address the two signatures:

$$\begin{aligned} pp \rightarrow t\bar{t} + E_{\mathrm {T}}^{\mathrm {miss}} \quad \text {and}\quad pp \rightarrow c\bar{c} + E_{\mathrm {T}}^{\mathrm {miss}}, \end{aligned}$$
(3)

where \(E_{\mathrm {T}}^{\mathrm {miss}} \) is the imbalance in transverse momentum in the event generated by the undetected neutralinos. Squark flavour mixing opens up a third final state,

$$\begin{aligned} pp \rightarrow tc + E_{\mathrm {T}}^{\mathrm {miss}}, \end{aligned}$$
(4)

where one squark decays into a top quark and the other one into a charm quark [29].

Fig. 1
figure 1

Dependance of the branching ratios BR(\(\tilde{u}_1\rightarrow t{\tilde{\chi }}^0_1\)) (dashed) and BR(\(\tilde{u}_1\rightarrow c{\tilde{\chi }}^0_1\)) (solid) on the squark mixing angle \(\theta _{tc}\) for various mass configurations. In the left panel, the squark mass is fixed to 500 GeV (red) and 1000 GeV (blue), with the neutralino mass being set to \(m_{\tilde{\chi }^0_1}=50\,{\mathrm {\ GeV}}\). In the right panel, the neutralino mass varies and is fixed to 50 GeV (red), 200 GeV (blue) and 300 GeV (cyan), for a squark mass of \(m_{\tilde{u}_1}=500\) GeV

In Fig. 1, we illustrate the \(\theta _{tc}\)-dependence of the squark branching ratios associated with the decays of Eq. (2). We observe that regardless the squark and neutralino mass configuration, there always exists a \(\theta _{tc}\) value for which both decay modes have 50% branching ratio, which means that half of the signal events would produce the final state of Eq. (4). Moreover, differences in the functional behaviour of the branching ratios for different mass hierarchies become only noticeable close to threshold, when the mass splitting between the decaying squark and the neutralino is small. This configuration is not considered further in this paper, as the phase space available for the decay is limited and the best experimental sensitivity is achieved with monojet or monotop probes [38].

2.2 Reinterpretation of current LHC Run 2 results

Fig. 2
figure 2

Reinterpretation, in the context of our simplified model, of the ATLAS bounds on top squarks obtained with single-leptonic probes [12] and on Supersymmetry when charm tagging is used [15]. The results are presented in the \((m_{\tilde{u}_1}, m_{\tilde{u}_2})\) plane (left) and \((m_{\tilde{u}_1}, \theta _{tc})\) plane (right), and the stars correspond to the official ATLAS results in the non-flavour-mixing case. The excluded region lies between the exclusion contour and the left-side of the figure

The ATLAS and CMS collaborations have performed several direct searches for top squarks, mostly in a setup where they are pair-produced and decay into a pair of top-antitop quarks and missing energy as indicated by the first equation of Eq. (3). With no observation of any hint for new physics, the most stringent constraints arise from LHC Run 2 analyses of proton-proton collisions at a centre-of-mass energy of 13 TeV [1,2,3,4,5,6,7,8,9,10,11,12]. All these searches lead to exclusion limits on the top squark mass of the order of 1 TeV. Bounds on first and second generation squarks are similar when one single light squark species is considered together with a decay into light jets and missing transverse energy, whereas they reach 1.5 TeV for models featuring four mass-degenerate first and second generation squarks [13, 14]. The most sensitive stop searches, yielding a similar expected sensitivity for low neutralino masses, are the ones addressing final states with either zero or one lepton. We therefore choose the recent ATLAS search for top squarks in final states with one lepton of Ref. [12] as a benchmark for getting conservative Run 2 constraints on our model.

Additionally, the ATLAS collaboration has carried out an analysis targeting top squarks decaying into charm and missing energy or charm squarks [15], based on the experimental tagging of jets produced from the fragmentation of charm quarks. As this signature is expected to play a significant role for getting handles on the considered squark inter-generational mixings, we use the analysis in Ref. [15] as a second LHC Run 2 benchmark to evaluate the existing constraints on our simplified model.

We perform a three-dimensional parameter space scan and vary independently the two squark masses (\(m_{\tilde{u}_1}\) and \(m_{\tilde{u}_2}\)), as well as the top-charm squark mixing angle \(\theta _{tc}\). As mentioned above, the neutralino mass has been fixed to 50 GeV, so that our results are valid as long as the squark masses are much larger than the neutralino mass. For each considered point, we evaluate the sensitivity of the two searches of Refs. [12, 15] and present the results in Fig. 2. The excluded region lies between the exclusion contour and the left-side of the figure. Concerning the stop analysis [12], we rely on the acceptances and efficiencies that have been officially provided by the ATLAS collaboration for each of the ‘discovery tN_med’ (targeting moderate stop masses) and ‘discovery tN_high’ (targeting high stop masses) regions. We then estimate the two corresponding signal yields (\(N_{\mathrm{sig}}\)), considering next-to-leading order (NLO) stop pair-production rates corrected by the resummation of the threshold logarithms at the next-to-leading logarithmic (NLL) accuracy [39] and the appropriate branching ratios. These signal yields are then compared to the ATLAS model-independent upper limit (\(N^{\mathrm{obs~limit}}_{\mathrm{non-SM}}\)) for each of the regions. If the ratio of these two yields exceeds one, the signal point is considered excluded. While providing acceptance and efficiency values only for the inclusive ‘signal’ regions, the ATLAS analysis employs a multi-bin fit in the most sensitive distribution for the final exclusion limit estimation. For this reason the recast exclusion contours presented in Fig. 2 represent a conservative estimate of the effective reach of the ATLAS search. We rely on the same procedure to extract the constraints from the charm-tagging analysis of Ref. [15].

In the left panel of Fig. 2, we consider a class of benchmark scenarios where the two squark eigenstates are maximal admixtures of the top and charm flavours (\(\theta _{ct} = \frac{\pi }{4}\)) and we vary the two masses independently (with \(m_{\tilde{u}_1}< m_{\tilde{u}_2}\)). The total new physics production rate is here solely driven by the lightest of the two states, except for the region where the mass splitting of the two squarks is small. For sufficiently high splittings, the exclusion is thus independent of \(m_{\tilde{u}_2}\), and squarks are found to be constrained to be heavier than about 550 GeV. Compared with the more standard MFV case where the two eigenstates are also flavour eigenstates (and where the bounds are of about 1 TeV), the limits are hence weakened by almost 500 GeV. The large value of the top-charm mixing angle indeed implies that the two signal regions of the stop analysis of Ref. [12], specifically targeting final states with the decay products of two top quarks, are less populated by virtue of the large decay fraction into charm jets BR\((\tilde{u}_1 \rightarrow c {\tilde{\chi }}^0_1)\). In the parameter space region defined by

$$\begin{aligned} m_{\tilde{u}_1}, m_{\tilde{u}_2} \lesssim 750~\mathrm{GeV}, \end{aligned}$$
(5)

the situation is somewhat different as the two squark mass eigenstates contribute to a potentially observable new physics signal. This partly compensates the loss due to the smaller branching ratio into tops, so that the obtained limits are stronger than when the second eigenstates is heavier. The charm-tagging analysis of Ref. [15] always implies weaker bounds for this specific classes of scenarios (the number of events populating the signal regions being very small), and the corresponding results are thus omitted.

In the right panel of Fig. 2, we reinterpret the ATLAS limits in the \((m_{{\tilde{u}}_1}, \theta _{tc})\) plane, i.e. we decouple the second eigenstate. Our results exhibit the complementary effect of the top-charm squark mixing angle on the bounds. For \(\theta _{tc}=0\), the lightest squark is purely of charm flavour, so that the ATLAS stop search is insensitive to the signal and the limits (\(m_{{\tilde{u}}_1} \gtrsim 800\) GeV) solely arise from the ATLAS charm-tagging analysis. With the mixing angle increasing, the \(c\bar{c}+E_{\mathrm {T}}^{\mathrm {miss}} \) production rate decreases so that the bounds are progressively weakened. On the other hand, the increase in \(\theta _{tc}\) implies that while the signal regions of the charm-tagging analysis are more and more depleted due to the lower and lower BR\((\tilde{u}_1 \rightarrow c{\tilde{\chi }}^0_1)\) branching ratio, the signal regions of the stop analysis are more and more populated due to the increasing BR\((\tilde{u}_1 \rightarrow t {\tilde{\chi }}^0_1)\) branching ratio. In the limit for which the lightest squark is purely of top flavour, its mass is constrained to be at least 825 GeV. In the maximal-mixing condition, the mass constraints for both analyses are below 600 GeV, which is the minimum mass value for which experimental acceptances are available for both the considered benchmark analyses.

We superimpose to the results of our recasting the official limits observed by ATLAS, represented by stars on the right panel of Fig. 2 for the cases where the lightest squark is of a definite flavour. The usage of multi-bin signal regions increases the limits by about 50–100 GeV.

3 Collider projections for the reach of the tc channel

We have shown in the previous section that the current experimental searches focusing on pair production of squarks that carry a well-defined flavour have a significantly reduced sensitivity to models with sizeable values of flavour mixing. The issue may be addressed by developing a dedicated analysis targeting the \(tc + E_{\mathrm {T}}^{\mathrm {miss}} \) channel, which has its maximum rate in the case of maximum mixing. We describe a possible implementation of such an analysis in this section. In particular, we will focus on the case in which the top quark decays semileptonically, resulting in a final state with an isolated lepton, a b-jet, a c-jet and missing transverse energy.

Our study assumes proton-proton collisions at a centre-of-mass energy of 14 TeV, and integrated luminosities of 300 and 3000 \(\mathrm{fb}^{-1}\), corresponding to the expected configurations for the coming LHC runs.

3.1 Monte Carlo simulation

In order to simulate our signal, we have implemented the model of Sect. 2.1 into Feynrules 2.0 [40] to get a UFO model [41] to be used within the MadGraph5_aMC@NLO framework [42]. We have generated leading-order (LO) hard-scattering matrix elements for squark pair-production and decay, that we have convoluted with the leading-order set of NNPDF 3.0 parton distribution functions [43]. Parton showering and hadronisation have been handled with Pythia 8.2 [44], and each event has been reweighted so that the corresponding total rate matches the production cross-section estimated at the NLO + NLL accuracy [39]. We generate a grid in the parameter space of the model, the lightest squark mass being varied in the [600 GeV, 1.5 TeV] window by steps of 100 GeV, and the neutralino mass in the [50 GeV, 900 GeV] window in steps of 50 GeV for \(m_{{{\tilde{\chi }}}^0_1}<400\, {\mathrm {\ GeV}}\) and of 100 GeV above. The squark mixing angle is fixed to \(\pi /4\).

As stated above, we focus on the signal topology with one isolated lepton (electron or muon), jets and missing transverse energy. The SM processes which can mimic this topology involve one or two leptons originating either from the decay of a W or a Z boson, or from leptonically-decaying tau leptons. We consequently generate events for SM \(t\bar{t}\), Wt, t-channel single top, \(t\bar{t}W\), \(t\bar{t}Z\), tWZ, tZ, \(W\)+jets, \(Z\)+jets, WW, WZ and ZZ production. For \(t\bar{t}\), single top and diboson processes, events are simulated at the NLO in QCD within the Powheg Box framework [45]. Samples for the remaining processes are then generated at LO, using MadGraph5_aMC@NLO. We consider matrix elements featuring a variable number of additional jets that we merge according to the CKKW prescription as implemented in Pythia 8 [46]. For \(W\)+jets and \(Z\)+jets, we merge samples describing final-states containing up to four additional partons, whereas for \(t\bar{t}W\) and \(t\bar{t}Z\) production, the matrix elements are allowed to include up to two extra partons. All those events are reweighted so that the total rates match the next-to-next-to-leading order (NNLO) cross-sections if available, or the NLO ones otherwise.

Jets are reconstructed according to the anti-\(k_T\) jet algorithm [47] with a jet radius parameter set to \(R = 0.4\), as implemented in FastJet [48]. Moreover, jets are labelled as b-jets if the angular distance \(\varDelta R\equiv (\varDelta \phi ^2+\varDelta \eta ^2)^{1/2}\) between the jet and the nearest B-hadron satisfies \(\varDelta R < 0.5\). Similarly, we define c-jets as jets that fail b-tagging and for which there exists a charmed hadron lying at an angular distance smaller than 0.5 from the jet. Any jet that is not identified as a b-jet or as a c-jet is labelled as a light jet. The missing transverse momentum \(\mathbf {p}^{\mathrm {\ miss}}_\mathrm {T} \), with magnitude \(E_{\mathrm {T}}^{\mathrm {miss}} \), is estimated by the vector sum of the transverse momenta of all invisible particles.

Detector effects are simulated by smearing the momenta of all reconstructed objects and by applying reconstruction efficiency factors in a way that reproduces the performance of the ATLAS detector [49, 50], as described in Ref. [51]. In particular, we include b-tagging and c-tagging efficiency and rejection factors based on the performance reported in Refs. [52, 53], and we adopt working points corresponding to an average b-tagging efficiency of \(\epsilon _b(b)=77\%\) for charm and light jet rejection factors of \(1/\epsilon _b(c)=4.5\) and \(1/\epsilon _b(l)=140\) respectively, and an average c-tagging efficiency of \(\epsilon _c(c)=30\%\) for rejection factors of \(1/\epsilon _c(b)=18\) and \(1/\epsilon _c(l)=5\) for b-jets and light jets respectively. Such a choice is aimed at optimising background rejection when it is dominated by final states featuring two b-jets. As there is currently no public information on the correlations between the b-tagging and c-tagging algorithms used by the collaborations, we do not allow a jet to be b-tagged and c-tagged simultaneously. We instead first select jet candidates based on their kinematics before applying either b-tagging or c-tagging.

We have compared our approach with an independent simulation based on the publicly available detector simulation software Delphes 3 [54], and have found good agreement between the two methods.

3.2 Event selection

The topology of interest includes one isolated lepton (electron or muon) arising from the top decay, jets including one b-jet (also issued from the top decay) and one c-jet, as well as missing transverse energy carried by the two neutralinos. Consequently, we preselect events by requiring the presence of exactly one isolated electron or muon with a transverse momentum \(p_{\mathrm {T}} >25 \hbox { GeV}\) and a pseudorapidity \(|\eta |<2.5\), and of at least one \(b\)-tagged jet with \(p_{\mathrm {T}} >50\, {\mathrm {\ GeV}}\) and \(|\eta |<2.5\). We moreover ask the invariant mass of at least one of the possible systems made of a b-jet and the lepton to fulfil \(m_{b\ell } < 160\, {\mathrm {\ GeV}}\), since in the signal case the lepton and the b-jet originate from a top decay so that \(m_{b\ell }\) is bounded to be smaller than around 153 GeV.

The dominant backgrounds at this point are comprised of \(t\bar{t} \) events with either one or both top quarks decaying leptonically, single top events and \(t\bar{t}Z \) events with an invisible Z-boson decay. As all backgrounds where the missing energy originates from a leptonic W-boson decay (\(W \rightarrow \ell \nu \)) feature

$$\begin{aligned} m_\mathrm {T}^{lep} \!\equiv \!\sqrt{2\,|\mathbf {p}^{\mathrm {\ \ell }}_\mathrm {T} |\,E_{\mathrm {T}}^{\mathrm {miss}} \,(1\!-\!\cos \varDelta \phi (\mathbf {p}^{\mathrm {\ \ell }}_\mathrm {T},\mathbf {p}^{\mathrm {\ miss}}_\mathrm {T}))}< m_W, \end{aligned}$$
(6)

we require \(m_\mathrm {T}^{lep} >160\hbox { GeV}\) to increase the signal over background ratio. In the definition of \(m_\mathrm {T}^{lep} \), \(\mathbf {p}^{\mathrm {\ \ell }}_\mathrm {T} \) is the lepton (vector) transverse momentum and \(\varDelta \phi (\mathbf {p}^{\mathrm {\ \ell }}_\mathrm {T},\mathbf {p}^{\mathrm {\ miss}}_\mathrm {T})\) the angle between \(\mathbf {p}^{\mathrm {\ \ell }}_\mathrm {T} \) and \(\mathbf {p}^{\mathrm {\ miss}}_\mathrm {T} \). Moreover, most of these backgrounds exhibit two b-jets in the final state, whereas the signal features in contrast one b-jet and one c-jet. Two strategies can therefore be envisaged to separate the signal from the backgrounds. Either one could veto the presence of any additional b-tagged jet besides the one required at the preselection level (Case-A), or one could enforce, in addition, the presence of an extra c-tagged jet (Case-B). From naive calculations based on the efficiencies of the different tagging algorithms, the signal over background ratio is improved by a factor of about 1.5 more for the Case-B strategy, but at the price of an overall reduction in statistics by a factor of approximately 3. Both approaches are thus pursued in the following. For the Case-A strategy, we veto the presence of any extra b-jet and impose that there is an extra light jet with \(p_{\mathrm {T}} >100\, {\mathrm {\ GeV}}\) failing b-tagging. In contrast, for the Case-B strategy, we require that only one b-tagged jet satisfies \(m_{b\ell } < 160\, {\mathrm {\ GeV}}\), we additionally impose that the leading jet fulfilling \(m_{j\ell } > 160\, {\mathrm {\ GeV}}\) is c-tagged and has a \(p_{\mathrm {T}} >100\) GeV, and we ask all remaining jets with \(m_{j\ell } > 160\, {\mathrm {\ GeV}}\) to fail b-tagging.

Fig. 3
figure 3

Distributions in the \(m_{\mathrm {T2}_{blj}}\) (left) and \(|\varDelta \phi _{\mathrm{min}}|\) (right) variables after imposing all cuts of the Case-A analysis strategy, excepted the one corresponding to the represented variable. The different background contributions and two representative signal scenarios are shown for an integrated luminosity of 300 fb\(^{-1}\). For the \(|\varDelta \phi _{\mathrm{min}}|\) distribution, the \(m_{\mathrm {T2}_{blj}}\) variable is required to be larger than 400 GeV

In order to further reduce the dileptonic \(t\bar{t}\) background where one of the leptons escapes identification, we make use of the now standard asymmetric \(m_{\mathrm {T2}}\) variable (denoted \(am_{\mathrm {T2}}\)) [55, 56] that consists in a variant of the \(m_{\mathrm {T2}}\) observable. The \(am_{\mathrm {T2}}\) variable is built from two legs (corresponding to the two decay chains) containing both a visible part and an invisible part, and it requires two test masses corresponding to the invisible mass attached with each leg. The visible part of the first leg is built using the sum of the momenta of the b-tagged jet and of the lepton, with a test mass that is set to zero. The visible part of the second leg is built from the remaining jet with the highest b-tagging weight and \(m_W\) is used as a test mass. The targeted background distribution featuring an end-point at approximately 160 GeV, we impose an \(am_{\mathrm {T2}} >200\, {\mathrm {\ GeV}}\) cut. In addition, the background can be further reduced by constructing another transverse \(m_{T2}\) variable. The signal topology features one squark leg where there is a hard c-jet, such that the distribution in the transverse mass built from the transverse momentum of the c-jet and the one of the neutralino exhibits an end-point at \((m_{{\tilde{u}}_1}^2-m_{{{\tilde{\chi }}}^0_1}^2)^ {1/2}\). This feature can be exploited by constructing an appropriate \(m_{\mathrm {T2}_{blj}}\) variable. The visible part of the first leg is built from the sum of the momenta of the b-tagged jet and of the lepton, together with a vanishing test mass. The visible part of the second leg uses the hardest non-b-tagged jet or the c-tagged jet for the Case-A and Case-B strategies respectively, and again a vanishing test mass. We impose a selection on \(m_{\mathrm {T2}_{blj}}\) depending on the squark-neutralino mass splitting in order to optimise the sensitivity to the signal. This optimisation is performed by varying the cut threshold from 300 to 600 GeV in steps of 50 GeV.

Finally, it is found that after all cuts, the missing transverse momentum is aligned with one of the jets of the event for the backgrounds, whereas there is no preferential direction for the signal. We therefore apply a cut on the minimum azimuthal angle separation between any jet and the missing transverse momentum, \(|\varDelta \phi _{\mathrm{min}}|>0.6\).

As an illustration, we present, in Fig. 3, the distribution in the \(m_{\mathrm {T2}_{blj}}\) (left) and \(|\varDelta \phi _{\mathrm{min}}|\) variable (right) for the different SM backgrounds and two representative signal benchmark points. All selection cuts from the Case-A analysis strategy are imposed, but the one on the represented variable. On the left panel, we can observe that a selection of \(m_{\mathrm {T2}_{blj}}\) \(> 400\) GeV is sufficient to separate the signal from the background for the lighter of the chosen benchmark models. On the right panel, we show instead the \(|\varDelta \phi _{\mathrm{min}}|\) distribution after including a cut of 400 GeV on \(m_{\mathrm {T2}_{blj}}\). We can again observe that a significant improvement of the signal to background ratio can be achieved by imposing a \(|\varDelta \phi _\mathrm{min}|>0.6\) cut.

4 Results

On the basis of the analysis strategy outlined in the previous section, we estimate the LHC sensitivity to supersymmetric scenarios featuring mixed stop-scharm states with 300 \(\mathrm{fb}^{-1}\) and 3000 \(\mathrm{fb}^{-1}\) of integrated luminosity. For the latter configuration, we assume no modification in the detector performances for the high-luminosity LHC. The sensitivity is extracted by means of a test statistics based on a profiled likelihood ratio, and we make use of the CLs method [57] to obtain 95% confidence level (CL) exclusion limits. The statistical analysis is performed with the RooStat toolkit [58] and we assume systematic uncertainties of 20% and 5% on the SM backgrounds and on the signal respectively. The results are presented in terms of the upper limits, at the 95% CL, on the ratio of the signal yields to the corresponding benchmark predictions, denoted as \(\sigma ^{\mathrm {excl}}/\sigma ^{\mathrm {SUSY}}\).

Fig. 4
figure 4

Left: sensitivity of the LHC to our mixed stop-scharm scenarios given as 95% CL exclusion contours in the \((m_{\tilde{u}_1}, m_{{\tilde{\chi }}^0_1})\) plane for the Case-A (red) and Case-B (blue) analysis strategies and for 300 fb\(^{-1}\) (solid) 3000 fb\(^{-1}\) (dashed). The projected excluded region lies between the exclusion contour and the bottom-left side of the figure. Right: signal over background ratio as a function of \(m_{\tilde{u}_1}\)for the Case-A (red, rounds) and Case-B (blue, squares) analysis strategies when one imposes that \(m_{\mathrm {T2}_{blj}} >550 \, {\mathrm {\ GeV}}\)

We show in the left panel of Fig. 4 the analysis reach in the \((m_{\tilde{u}_1},m_{{\tilde{\chi }}^0_1})\) plane both for the Case-A (red) and Case-B (blue) analysis strategies and for 300 fb\(^{-1}\) (solid) and 3000 fb\(^{-1}\) (dashed) of integrated luminosity. The region that lies between the exclusion contour and the bottom-left side of the figure will be excluded at the future runs of LHC. The expected 95% upper limit on \(m_{\tilde{u}_1}\)for \(m_{{\tilde{\chi }}^0_1}=50\, {\mathrm {\ GeV}}\) is \(1050 {\mathrm {\ GeV}}\) for Case-A, and \(920 {\mathrm {\ GeV}}\) for Case-B for an integrated luminosity of 300 fb\(^{-1}\). The large difference is due to the fact that the analysis reach is in this case dominated by statistics, which is lower for the analysis based on c-tagging, due to the 30% efficiency of the chosen c-tagging working point. For an integrated luminosity of 3 ab\(^{-1}\) the difference is reduced, with a reach of \(1280 {\mathrm {\ GeV}}\) for Case-A and \(1240 {\mathrm {\ GeV}}\) for Case-B.

In order to better understand the relative performance of the two analysis strategies, we present in the right panel of Fig. 4 the dependence of the signal over background ratio (S / B) on the squark mass \(m_{\tilde{u}_1}\) for \(m_{{\tilde{\chi }}^0_1}=50\, {\mathrm {\ GeV}}\), \(\theta _{tc}=\pi /4\) and when a \(m_{\mathrm {T2}_{blj}}\) > 550 GeV cut is applied. As expected, the S / B ratio is higher when c-tagging is incorporated. Comparisons of results stemming from analyses with and without c-tagging, or relying on different c-tagging working points could be used to get information on the flavour content of the observed squark, which is the main information one would like to extract in case of a discovery. In the Case-B analysis, we have chosen a c-tagging working point which optimises S / B, but with a similar efficiency for c-jets and light jets, and thus not ideal for discriminating the flavour of the signal. A different c-tagging algorithm working point featuring a very high rejection for light jets, as e.g. in Ref. [15] with \(\epsilon _c(c)=18\%\), \(1/\epsilon _c(b)= 20\) and \(1/\epsilon _c(l)=200\) would yield a lower overall sensitivity, but might be used to discriminate between different flavour mixing hypotheses for the signal.

In Fig. 5, we show the 95% CL exclusion limits in the \((m_{{\tilde{u}}_1}, \theta _{tc})\) plane for a fixed neutralino mass of 50 GeV. Recasts of the 13 TeV exclusion limits obtained by the ATLAS experiment with 36 \(\mathrm{fb}^{-1}\) are in addition overlaid (see Sec. 2.2), the blue curve corresponding to the ATLAS search for stop in the singly-leptonic mode [12] and the red one to the ATLAS search for squarks based on charm-tagging [15]. The region between the recasted exclusion contour and the top-left (bottom-left) side of the figure is excluded by the \(t\bar{t}\) + \(E_{\mathrm {T}}^{\mathrm {miss}}\) (\(c\bar{c}\) + \(E_{\mathrm {T}}^{\mathrm {miss}}\)) analysis. The exclusion limits expected from the tc + \(E_{\mathrm {T}}^{\mathrm {miss}}\) Case-A analysis strategy are shown as solid and dotted black lines for integrated luminosities of 300 \(\mathrm{fb}^{-1}\) and 3000 \(\mathrm{fb}^{-1}\) respectively. Moreover, we also include the expectation of such an analysis at 14 TeV, with 36 \(\mathrm{fb}^{-1}\) of luminosity. For tc + \(E_{\mathrm {T}}^{\mathrm {miss}}\) analysis, the projected excluded region lies between the exclusion contour (solid, dashed and dotted black lines) and the left side of the figure. This figure clearly illustrates the strength of the tc + \(E_{\mathrm {T}}^{\mathrm {miss}}\) analysis we are proposing, covering a region of the parameter space not accessible with current searches relying on the MFV paradigm.

Fig. 5
figure 5

Present and expected exclusion limits in the \((\theta _{tc}, m_{{\tilde{u}}_1})\) plane. The area between the recasted exclusion contour and the top-left (bottom-left) side of the figure is excluded by the \(t\bar{t}\) + \(E_{\mathrm {T}}^{\mathrm {miss}}\) (\(c\bar{c}\) + \(E_{\mathrm {T}}^{\mathrm {miss}}\)) analysis. The region that lies between the exclusion contour (solid, dashed and dotted black lines) and the left side of the figure will be excluded at the future runs of LHC using tc + \(E_{\mathrm {T}}^{\mathrm {miss}}\) analysis. See the text for details

5 Summary

We have studied a supersymmetric scenario departing from the traditional MFV paradigm. Flavour mixing including squarks of the first generation and left-handed partners being strongly constrained by data, we focus on specific scenarios where right-handed stop and scharm mix. In this context, squark pair production and decay can yield three distinctive signatures, namely top pairs plus missing energy, charm-jet pairs with missing energy, and tc plus missing energy.

By recasting existing LHC searches for top and charm squarks, we have shown that a significant fraction of the parameter space is evading present LHC constraints and that mixed squark states with masses around 600 GeV remain a phenomenologically viable option. This is in particular true when the top-charm mixing is maximal, as both \(t\bar{t}\) + \(E_{\mathrm {T}}^{\mathrm {miss}}\) and \(c\bar{c}\) + \(E_{\mathrm {T}}^{\mathrm {miss}}\) signals are weakened. This apparent loophole in the experimental searches could be filled by the design of a dedicated tc + \(E_{\mathrm {T}}^{\mathrm {miss}}\) search for squarks.

To this purpose, we have developed two search strategies targeting the production of squarks containing two flavour components, both involving leptons, b-tagged jet and \(E_{\mathrm {T}}^{\mathrm {miss}}\). The core difference between them consists in using charm-tagging in one of them, making hence use of the presence of a charm jet in the final state as a handle on new physics. With charm-tagging efficiencies presently achieved by experiments being lower than for b-tagging, the requirement of a charm-tagged jet implies a loss of signal statistics and hence of mass reach for an integrated luminosity of 300 \(\mathrm{fb}^{-1}\). The statistical limitation on the signal are however less important for the high-luminosity phase of the LHC, where mixed squarks as heavy as 1.3 TeV could be targeted for both analysis strategies. The advantage of pursuing several analyses in parallel is that the comparison of the results from analyses relying on different requirement and/or working points for the flavour tagging of jets opens the door to probing the squark flavour structure and hence allows the potential observation of departures from the MFV paradigm.

As charm tagging is being actively developed by both the ATLAS and CMS collaborations, we urge them to add to their search program a dedicated tc + \(E_{\mathrm {T}}^{\mathrm {miss}}\) analysis that could provide sensitivity to new physics scenarios where the top partners are not flavour eigenstates.