1 Introduction

The next large collider facility could be a high-energy electron–positron collider. Linear \(e^+e^-\) colliders are the best tool to explore the energy range from several 100 GeV to a few TeV. The Technical Design Report of the International Linear Collider (ILC [1, 2]) project envisages a programme of precision Higgs and top physics at centre-of-mass energies \(\sqrt{s}= 250\), 500 GeV and, after an energy upgrade, 1 TeV. The compact linear collider (CLIC [3]) scheme has been shown to reach accelerating gradients that extend the \(e^+e^-\) programme into the multi-TeV regime. CLIC envisages a first phase at \(\sqrt{s}= 380\) GeV, followed by its high-energy programme of multi-TeV operation, with stages at 1.5 and 3 TeV [4]. A large circular machine, as envisaged by the FCCee [5] and CEPC [6] projects, could provide high luminosity at 250 GeV. FCCee may reach the top-quark pair production threshold [5]. On a longer time scale a muon collider [7] also has the potential to reach the multi-TeV regime.

Measurements of hadronic final states are a key ingredient of the programme of any next-generation lepton collider. Excellent jet reconstruction is essential to characterise the couplings of the Higgs boson and top quark at the sub-percent level. To distinguish hadronic W and Z boson decays a jet energy resolution of approximately 3% is required. The linear collider detector concepts [8, 9] achieve this performance with highly granular calorimeters [10, 11] and particle-flow algorithms [12]. Excellent jet clustering is needed to benefit fully from their potential.

The increase in energy comes with a number of challenges for jet reconstruction. Compared to LEP and SLC, high-energy machines produce an abundance of multi-jet final states, final states with multiple energy scales (in associated production), forward-peaked processes and highly boosted objects. Backgrounds such as \(\gamma \gamma \rightarrow \) hadrons production are increasingly important at high energy [13]. The classical \(e^+e^-\) algorithms cannot cope with this environment, in particular with the beam-induced background [3, 14, 15]. A critical evaluation of jet reconstruction at lepton colliders is therefore mandatory.

In this paper we study the performance of jet reconstruction in multi-TeV \(e^+e^-\) collisions in detail. This work focuses on exclusive reconstruction, where the number of jets is specified in agreement with the number of partons in the target final state, as it has been shown to yield superior results in previous \(e^+e^-\) benchmark analyses [3, 15]. This choice excludes cone algorithms (such as SiSCone [16]) from consideration, as well as algorithms with a purely angular distance criterion (i.e. Cambridge [17]), and anti-\(k_{t}\) type algorithms, which are limited to inclusive clustering.

We benchmark the performance of a number of sequential recombination algorithms: the classical \(e^+e^-\) \(k_\mathrm {t}\) algorithm [18] used by the LEP experiments and SLD, the longitudinally invariant \(k_{\mathrm {t}}\) algorithm [19, 20] used at hadron colliders and a generalisation of the Valencia algorithm [21], a robust \(e^+e^-\) algorithm.

Several aspects of jet reconstruction performance are studied in simulated events. In a particle-level study we estimate the size of perturbative and non-perturbative corrections to the jet energy. We establish their dependence on the process and the centre-of-mass energy, and on the parameters of the jet algorithms. These particle-level simulations are also used to study the impact of energy deposited on the signal event by background processes. We conclude by using a realistic simulation of the CLIC detector and particle-flow reconstruction to study the jet reconstruction performance in top-quark pair production and di-Higgs boson production.

The layout of this paper is as follows. In Sect. 2 we review the challenges of jet reconstruction at high-energy lepton colliders. In Sect. 3 the jet reconstruction algorithms are introduced. In Sect. 4 the perturbative and non-perturbative corrections are studied. In Sect. 5 we study the response of the algorithms to the signal jet and the background. Section 6 presents the results of a full-simulation study of a few benchmark processes. In Sect. 7 we discuss possible directions for future work and in Sect. 8 the most important findings of this work are summarised.

2 Challenges for jet reconstruction at high-energy \(e^+e^-\) colliders

The experimental environment at previous lepton colliders, such as LEP and SLC, was very benign compared to that at hadron colliders. While this remains true at future high-energy lepton machines, jet reconstruction faces a number of new challenges.

2.1 Multi-jet final states

Future lepton colliders offer the possibility to study \(2 \rightarrow 4\), \(2 \rightarrow 6\), and even \( 2 \rightarrow 8\) processes. The dominant branching ratios of the W, Z and Higgs bosons are from hadronic decays. Final states with four jets, most notably \(e^+ e^- \rightarrow Z h\), play a key role in the physics programme of any future electron–positron collider. At high-energy final states with six jets (e.g. \(e^+ e^- \rightarrow t\bar{t} \)), or even eight or ten jets (e.g. \(e^+ e^- \rightarrow t\bar{t} h\)), become important. Imperfect clustering of final-state particles can affect the reconstruction of hadronic decays in an important way.

The impact of incorrect assignments of final-state particles to jets is illustrated with an example. We consider the Higgs-strahlung process with hadronic decays of Z- and Higgs bosons, where reliable reconstruction of Z and h-boson candidates is the key to a precise measurement of the Higgs boson couplings [22]. The hard scattering process \(e^+ e^- \rightarrow Z h\) is simulated with the MadGraph_aMC@NLO [23] package and the \(Z \rightarrow q \bar{q}\) and \(h \rightarrow b \bar{b}\) decays and subsequent hadronisation with Pythia8 [24]. The Higgs boson mass in the simulation is 125 GeV. No beam energy spread, initial-state radiation, or modelling of background or the detector are included. Stable particles are clustered into jets with the Durham algorithm (exclusive clustering with \(N=\) 4).

Higgs boson candidates are formed by adding the four-vectors of the two jets that yield the di-jet mass closest to the Higgs boson mass. In Fig. 1 the invariant mass distribution of the Higgs boson candidates is shown for three centre-of-mass energies. We find that the distribution has a non-zero width, even in this relatively perfect simulation. The finite resolution is purely due to imperfect clustering of final-state particles into jets. The effect of confusion in jet clustering is less pronounced at higher centre-of-mass energy, as the greater boost of the Z- and Higgs bosons leads to a cleaner separation of the jets.

Fig. 1
figure 1

The reconstructed Higgs boson mass peak in \(e^+ e^- \rightarrow Z h\), \(Z \rightarrow q \bar{q}\), \(h \rightarrow b \bar{b}\) events. The three histograms correspond to centre-of-mass energies of 250 GeV (red, continuous line), 350 GeV (blue, dashed line) and 500 GeV (black, continuous line)

The most important challenge stems from the larger jet multiplicity. In events with only two jets (i.e. \(e^+ e^- \rightarrow Z h\) events with \(Z \rightarrow \nu \bar{\nu }\) and \(h \rightarrow b \bar{b}\)) the clustering contribution to the mass resolution is negligible. In the ILC and CLIC analyses of di-Higgs boson [25] and \(t\bar{t}h\) production [26, 27] jet clustering is found to be the limiting factor for the Higgs mass resolution.

Finally, we note that as the centre-of-mass energy increases, t-channel processes become more important. The most obvious example is vector-boson-fusion production of the Higgs boson. The final-state products at high energy are strongly forward-peaked [28] and special care is needed to ensure robust jet reconstruction performance over the full polar angle coverage of the experiment. Di-Higgs boson production through vector-boson fusion (\(e^+ e^- \rightarrow \nu \bar{\nu } hh\)) presents a double challenge of high jet multiplicity and forward jets. We therefore take this analysis at \(\sqrt{s} = \) 3 TeV as a benchmark.

2.2 Jet substructure

The production of very energetic gauge bosons, Higgs bosons and top quarks with hadronic decays, collectively denoted as boosted objects, poses a challenge to the experiments at a future high-energy collider. Whenever the energy of the decaying object exceeds its mass m significantly the highly collimated decay products typically cannot be resolved. In such cases, an analysis of the internal structure of jets at a scaleFootnote 1 \(R < 2 m/p_{\mathrm {T}}\) is performed to identify the boosted object [29, 30].

At high-energy linear colliders, the separation of boosted W- and Z-bosons and the reconstruction of boosted top quarks challenge the detector and reconstruction algorithms. The highly granular calorimeters [10, 11] of the Linear Collider detector concepts and the particle-flow paradigm are eminently suited for substructure analyses. We analyse the large-R jet mass resolution in top-quark pair production at \(\sqrt{s} =\) 3 TeV, as a first exploration of the jet substructure performance of experiments at future lepton colliders.

2.3 Beam-induced background

While the environment at lepton colliders remains much more benign than the pile-up conditions of high-energy hadron colliders, several background sources cannot be ignored in the detector design and evaluation of the performance of the linear collider experiments. The most relevant background source for jet reconstruction at linear \(e^+e^-\) colliders is \(\gamma \gamma \rightarrow \) hadrons production [3]: photons emitted from the incoming electron and positron beams (bremsstrahlung and beamstrahlung) collide and produce mini-jets of hadrons.Footnote 2 In high-energy colliders the probability to produce a mini-jet event in a given bunch crossing is of the order of one.

To evaluate the detector performance, ILC and CLIC detector concepts superpose a number of \(\gamma \gamma \rightarrow \) hadrons background events on the hard scattering process. The distribution of the particles formed in \(\gamma \gamma \) collisions is forward-peaked, with approximately constant density per unit of rapidity over the instrumented region (a feature also present in the pile-up due to minimum-bias events in proton–proton collisions). For CLIC at 3 TeV approximately 90% of the energy is deposited in the endcap calorimeters and only 10% in the central (barrel) regions of the experiment [3].

The impact of the background on the performance depends on the bunch structure of the accelerator and the read-out speed of the detector systems. In particular for machines based on radio-frequency cavities operated at room temperature the bunch spacing can be very small; CLIC envisages a bunch spacing of 0.5 ns. Background processes deposit 19 TeV in the detectors during a complete bunch train of 312 consecutive bunch crossings at \(\sqrt{s}=\) 3 TeV [3]. A selection based on the time stamp and transverse momentum of the reconstructed particles reduces this background to approximately 100 GeV on each reconstructed physics event. The relatively large bunch spacing of approximately 500 ns at the ILC allows the detector to distinguish individual bunch crossings. In our full-simulation study simulated \(\gamma \gamma \rightarrow \) hadrons events are overlaid on the signal events.

2.4 Initial-state radiation

In \(e^+e^-\) annihilation processes the system formed by the final-state products is, to first approximation, produced at rest in the laboratory. Initial-state radiation (ISR) photons emitted by the incoming electrons and positrons changes this picture somewhat. To estimate the magnitude of the boost introduced by ISR we generate events using a parton-level calculationFootnote 3 in MadGraph_aMC@NLO [23]. For each 2 \(\rightarrow \) 2 process \(e^+e^- \rightarrow XY\) we include also the 2 \(\rightarrow \) 3 process \(e^+e^- \rightarrow XY \gamma \). In Fig. 2 the fraction of the energy carried by the XY system is shown for several 2 \(\rightarrow \) 2 processes and for several centre-of-mass energies. For most s-channel processes, the ISR photon energy spectrum falls off very rapidly and the visible energy distribution displays a sharp peak at 1.

Fig. 2
figure 2

The fraction of the visible energy (the energy carried by all final-state products except the photon) in several processes, where a photon radiated off the initial- or final-state particles may accompany the final state products. The distributions correspond to pair production of light fermions in association with a photon (\(e^+e^- \rightarrow f \bar{f} (\gamma )\)), W-boson and top-quark pair production (\(e^+e^- \rightarrow W^+W^-(\gamma )\), \(e^+e^- \rightarrow t \bar{t} (\gamma )\)) and the Higgs-strahlung process \(e^+e^- \rightarrow Zh (\gamma )\). The centre-of-mass energy is indicated on the figure for each process

The boost of the system along the z-axis due to ISR remains relatively small close to the production threshold: for \(e^+e^- \rightarrow Zh (\gamma )\) at \(\sqrt{s} =\) 250 GeV and \(e^+e^- \rightarrow t \bar{t} (\gamma )\) at 500 GeV \(\beta _z = v_z/c\) is smaller than 0.1 in over 95 and 90% of the events, respectively.Footnote 4 For processes with a cross section that grows with \(\sqrt{s}\) the peak is even narrower. The role of ISR is only significant for radiative return to the Z in the process \(e^+e^- \rightarrow f \bar{f} (\gamma )\), where f is any fermion with mass less than half that of the Z boson.

At linear colliders with very narrow beams the luminosity spectrum displays a sizeable tail towards lower centre-of-mass energy [32]. Beam energy spread and beamstrahlung may therefore lead to a pronounced boost of the visible final-state objects in a small fraction of events. This effect is included in the full-simulation study of Sect. 6.

3 Jet reconstruction algorithms

In this section the jet algorithms considered in this paper are introduced. We discuss the most important differences and their implications for the performance. Three classes of sequential recombination algorithms are considered here:

  • The classical \(e^+e^-\) algorithms [18] and their generalisation with beam jets [33].

  • The longitudinally invariant algorithms developed for hadron colliders [19, 20].

  • The Valencia algorithm proposed in a previous publication [21], which is further generalised here.

3.1 The VLC algorithm

In Ref. [21] a robust jet reconstruction algorithm was proposed that maintains a Durham-like distance criterion based on energy and polar angle. It achieves a background resilience that can compete with the longitudinally invariant \(k_t\) algorithm. Here, we further generalise the definition of the algorithm.

The VLC algorithm has the following inter-particle distance:

$$\begin{aligned} d_{ij} = 2 \min {(E_i^{2 \beta },E_j^{2 \beta })} (1 - \cos {\theta _{ij}})/R^2, \end{aligned}$$
(1)

where R is the radius or resolution parameter. For \(\beta = \) 1 the distance is given by the transverse momentum squared of the softer of the two particles relative to the harder one, as in the Durham algorithm.

The beam distance of the algorithm is

$$\begin{aligned} d_{i\mathrm {B}} = E_i^{2\beta } \sin ^{2\gamma }{\theta _{i\mathrm {B}}}, \end{aligned}$$
(2)

where \(\theta _{i\mathrm {B}}\) is the angle with respect to the beam axis, i.e. the polar angle.

The two parameters \(\beta \) and \(\gamma \) allow independent control of the clustering order and the background resilience.Footnote 5 The \(\beta \) and \(\gamma \) parameters are real numbers that can take any value. For \(\beta =\gamma =1\) the expression for the beam distance simplifies to \(d_{i\mathrm {B}} = E^2 \sin ^2{\theta _{i\mathrm {B}}} = p^2_{\mathrm {T}i}\). We discuss the impact of different choices in Sect. 3.3.

This new version of the algorithm fulfils the standard IR-safety tests of the FastJet team. The VLC algorithm is available as a plug-in for the FastJet [33, 34] package. The code can be obtained from the “contrib” area [35].

3.2 Comparison of the distance criteria

The generalised distance criteria for three families of algorithms are summarised in Table 1.

Table 1 Summary of the distance criteria used in sequential recombination algorithms. Generalised inter-particle and beam distances are given for three main classes of algorithms: the classical \(e^+e^-\) algorithms (comprising a version with beam jets of the Cambridge [17] and Durham algorithms), the longitudinally invariant algorithms used at hadron colliders, which comprise longitudinally invariant \(k_t\), Cambridge–Aachen and anti-\(k_t\) and the robust \(e^+e^-\) algorithms introduced in Sect. 3.1

For all algorithms the clustering order can be modified by an appropriate choice of the n in the exponent of the energy (or \(p_{\mathrm {T}}\)) in the inter-particle distance (\(\beta \) in the VLC algorithm). The Durham (or \(k_t\)) algorithm, with \(n=1\), clusters pairs of particles starting with those that are soft and collinear (i.e. the inverse of the virtuality-ordered emission during the parton shower). Choosing \(n=\) 0 yields the Cambridge/Aachen algorithm, which has a purely angular distance criterion. The anti-\(k_t\) algorithm has \(n=-\)1.

In the generalised algorithms the area of jets is limited. Any particle with a beam distance [36] smaller than the distance to any other particle is associated with the beam jet, and therefore not considered part of the visible final state. This modification renders jet reconstruction very resilient to backgrounds. The radius parameter R governs the relative size of the inter-particle and beam distance and thus determines the size of the jet. In practice, the choice of R is a compromise between the background resilience of small-radius jets and the wish to capture the signal energy flow as fully as possible, which drives the choice of R to higher values. In the studies in the following sections, the R parameter is varied and the value that optimises the performance is retained. Compared to inclusive clustering at hadron colliders, where radius parameters of 0.4–0.5 are typical, an optimisation of the performance of exclusive clustering in the CLIC environment prefers much larger values, typically in the interval 1–1.5.

The generalised \(e^+e^-\) algorithm and the VLC algorithm have virtually the same inter-particle distance. However, the radius parameter R is redefined: the inter-particle distance \(d_{ij}\) denominator is \(R^2\) instead of \( 1 - \cos {R}\). The hadron collider algorithms replace the particle energy \(E_i\) and angle \(\theta _{ij}\) with quantities that are invariant under boosts along the beam axis, the transverse momentum \(p_{\mathrm {T}i}\) and the distance \(\Delta R_{ij} = \sqrt{(\Delta \phi )^2 + (\Delta y)^2}\), where \(\phi \) is the azimuthal angle in the usual cylindrical coordinates and y denotes the rapidity.

Fig. 3
figure 3

The area or footprint of jets reconstructed with \(R=\) 0.5 with the three major families of sequential recombination algorithms. The two shaded areas in each column correspond to a jet in the central detector (\(\theta = \pi /2\)) and to a forward jet (\(\theta = 7\pi /8\)). The jet axis is indicated with a cross

Detailed studies [3, 14] show that the longitudinally invariant \(k_t\) algorithm is much more resilient to the \(\gamma \gamma \rightarrow \) hadrons background than the classical and generalised \(e^+e^-\) algorithms.

For each of the algorithms the catchment area can be defined as the area where the distance \(d_{ij}\) between a soft, peripheral particle and the hard, central core of the jet is smaller than the beam distance. This definition corresponds to the passive area of Ref. [37]. The areas of a single central and forward jet with \(n=\)1 and \(R=\) 0.5 are indicated in Fig. 3. The footprint of the central jet (at \(\theta =\pi /2\)) is approximately circular for all algorithms. The area of the jet in the forward detector (at \(\theta =7\pi /8\)) shrinks considerably for the longitudinally invariant algorithms and the VLC algorithm. The reduced exposure in this region, where backgrounds are most pronounced, is the crucial feature for the enhanced resilience of these algorithms.

An analytical understanding of this property can be obtained by considering two test particles with energies \(E_i\) and \(E_j\) and separated by a fixed angle \(\Omega _{ij}\). For the generalised \(e^+e^-\) algorithms, both the distance \(d_{ij}\) between the two particles and the ratio \(d_{ij}/d_{i\mathrm {B}}\) of the inter-particle distance and the beam distance are independent of polar and azimuthal angle. For the longitudinally invariant algorithms the ratio \(d_{ij}/d_{i\mathrm {B}}\) increases (while the inter-particle distance decreases) as the two-particle system is rotated into the forward region. Finally, for the VLC algorithm the ratio \(d_{ij}/d_{i\mathrm {B}}\) increases as \(1/\sin ^{2\gamma } \theta \) in the forward region, with a slope that is similar to that of the longitudinally invariant algorithms for \(\gamma =\) 1. The distance \(d_{ij}\) is constant, as in classical \(e^+e^-\) algorithms.

A closer comparison of the shape of the footprint of the longitudinally invariant algorithms and the VLC algorithm show that, given identical jet axes, the former extend further into the forward region. This causes a slight difference in background resilience of both classes of algorithms.

3.3 Interpolation between algorithms

The two parameters \(\beta \) and \(\gamma \) of the VLC algorithm allow one to tailor the algorithm to a specific application. As these parameters are real numbers, one can interpolate smoothly between different clustering schemes.

The \(\beta \)-parameter that exponentiates the energy in inter-particle and beam distance governs the clustering order (similar to the exponent n in the generalised \(k_t\) algorithm). For \(\beta = 1\) clustering starts with soft, collinear radiation. Choosing \(\beta =0\) yields purely angular clustering, while \(\beta =-1\) corresponds to clustering starting from hard, collinear radiation. These integer choices of \(\beta \) correspond to \(k_t\), Cambridge/Aachen and anti-\(k_t\) clustering, respectively. Non-integer values of \(\beta \) interpolate smoothly between these three schemes.

The parameter \(\gamma \) in the exponent of the beam distance of the VLC algorithm provides a handle to control the shrinking of the jet catchment area in the forward regions of the experiment. After setting the R-parameter to the optimal value for central jets, the area of forward jets can be tuned by the choice of \(\gamma \) to ensure the required background resilience.

Fig. 4
figure 4

Diagram of the parameter space spanned by exponents \(\beta \) and \(\gamma \) of the VLC algorithm. On the y-axis generalisations with beam jets of the LEP/SLD algorithms are found, with the Cambridge algorithm with angular ordering at the origin and the Durham or \(k_t\) algorithm at \(\beta =1\). Choosing \(\beta =\) -1 yields reverses the clustering order (like in anti-\(k_{\mathrm {t}}\) algorithm [38]). Choosing non-zero and positive values for \(\gamma \) yields robust algorithms with a shrinking jet area in the forward region

We have seen that \(\gamma = \) 1 yields forward jets with a similar size of those of the longitudinally invariant algorithms for hadron colliders.Footnote 6 Values of \(\gamma \) greater than 1 further enhance the rise of the \(\frac{d_{ij}}{d_{i\mathrm {B}}}\) ratio in the forward region, causing the jet footprint to shrink faster. Values between 0 and 1 yield a slower decrease of the area when the polar angle goes to 0 or \(\pi \).

For \(\gamma =\) 0, \(d_{i\mathrm {B}} = E_{i}^{2\beta }\) and we retrieve the generalised \(e^+e^-\) algorithms with constant angular opening: the generalised Cambridge algorithm [17] for \(\beta =\) 0 and generalised \(k_{\mathrm {t}}\) or Durham [18] for \(\beta =\) 1. Choosing \(\beta =\) -1 yields an \(e^+e^-\) variant of the anti-\(k_{\mathrm {t}}\) algorithm [38]. A schematic overview of the algorithms in \((\beta , \gamma )\) space is given in Fig. 4.

4 Jet energy corrections

Before we turn to a detailed simulation including overlaid backgrounds and a model for the detector response, we study the perturbative and non-perturbative jet energy corrections of the algorithms. Both types of corrections are closely connected to the jet area [39]. In this section we quantify their impact, following the analysis of Ref. [39]. This first exploration of the stability of the algorithms should be extended in future work to quantify the impact of next-to-leading correction, as performed for instance in Ref. [40]. Also the robustness of the conclusions for a variety of different sets of parameters (tunes) of the Monte Carlo simulation merits further study.

4.1 Monte Carlo setup

The Monte Carlo simulation chain uses the MadGraph5_aMC@NLO package [23] to generate the matrix elements of the hard scattering \(2 \rightarrow 2\) event. Several processes are studied, but results in this Section focus on \(e^+e^- \rightarrow q\bar{q}\) at \(\sqrt{s}=\) 250 GeV and \(e^+e^- \rightarrow t\bar{t}\) with fully hadronic top decays at \(\sqrt{s}=\) 3 TeV. The four-vectors of the outgoing quarks are fed into Pythia 8.180 [24], with the default tune to LEP data, that performs the simulation of top-quark and W boson decays, the parton shower and hadronisation. No detector simulation is performed and initial-state radiation and beam energy spread are not included in the simulation. Particles or partons from the Pythia event record are clustered using FastJet 3.0.6 [33] exclusive clustering with \(N=\) 2. The default (“E-scheme”) recombination algorithm is used to merge (pseudo-) jets.

Fig. 5
figure 5

Jet energy correction as a function of the jet radius parameter R in \(e^+ e^-\rightarrow q \bar{q}\) production at \(\sqrt{s} =\) 250 GeV(left panel) and \(e^+ e^-\rightarrow t \bar{t}\) production at \(\sqrt{s}=\) 3 TeV(right panel). The continuous line corresponds to the median relative correction, the dashed line to the mean. Results are shown for three algorithms: the generalised \(e^+e^-\) algorithm, the longitudinally invariant \(k_t\) algorithm and the VLC algorithm with \(\beta = \) 1. The statistical uncertainties of the results are negligible and are not indicated

4.2 Definition of response and resolution

The jet energy and mass distributions often display substantial non-Gaussian tails and the choice of robust estimators has non-trivial implications. To estimate the centre of the distribution (i.e. the response) the mean and median are used. In some cases we present both, to give an indication for the skewness of the distribution. The width of the distribution (resolution) is estimated using the inter-quantile range \(\mathrm {IQR}_{34}\), which measures half the width of the interval centered on the median that contains 68% of all jets. We also use \(\mathrm {RMS}_{90}\), the root-mean-square of the values after discarding 5% outliers in both the low and high tails of the distribution.

4.3 Perturbative corrections

Following Ref. [39] we estimate the total energy correction by comparing the parton from the hard scatter to the jet of stable particles. For jets of finite size this correction is dominated by energy that leaks out of the jet. We indeed find that the distribution of the difference of parton and jet energy is asymmetric, with a long tail towards negative corrections, where the parton energy is larger than the energy captured in the jet. This energy leakage is most pronounced for jets with a small radius parameter, as expected.

In Fig. 5 the average (dashed line) and median (continuous line) relative energy correction are presented. The left plot corresponds to \(e^+ e^-\rightarrow q \bar{q}\) collisions at relatively low energy (\(\sqrt{s}= \) 250 GeV), while the right plot corresponds to \(e^+ e^-\rightarrow t \bar{t}\) at \(\sqrt{s}= \) 3 TeV. At a quantitative level the results show some dependence on the process, centre-of-mass energy and the generator tune for which they are obtained, but qualitatively the same pattern emerges in all cases. The energy correction decreases as the catchment area of the jet increases.

The energy corrections for the generalised \(e^+ e^-\) algorithm vanish relatively rapidly, with the median correction reaching sub-% level for \(R \sim 1\). The VLC and longitudinally invariant \(k_t\) algorithm show much slower convergence towards zero correction. This is entirely due to jets close to the beam axis. For central jets the three classes of algorithms yield identical results (within the statistical accuracy). The VLC and \(k_t\) algorithms have similar footprints and, indeed, very similar energy corrections.

The clustering order (as controlled by n in the generalised algorithm and by \(\beta \) in the VLC algorithm has a minor impact on the energy corrections. The (inclusive) Cambridge/Aachen algorithm and anti-\(k_t\) algorithm give similar results to the \(k_t\) variants of the same algorithm shown here.

Fig. 6
figure 6

Non-perturbative jet energy corrections to the jet energy as a function of the jet radius parameter R in \(e^+ e^-\rightarrow q \bar{q}\) production at \(\sqrt{s} =\) 250 GeV (left panel) and \(e^+ e^-\rightarrow t \bar{t}\) production at \(\sqrt{s}=\) 3 TeV (right panel). The continuous line corresponds to the median relative correction, the dashed line to the mean. Results are shown for three algorithms: the generalised \(e^+e^-\) algorithm, the longitudinally invariant \(k_t\) algorithm and the VLC algorithm with \(\beta = \) 1. The statistical uncertainties of the results are negligible and are not indicated

Fig. 7
figure 7

Non-perturbative corrections to the jet mass as a function of the jet radius parameter R in \(e^+ e^-\rightarrow q \bar{q}\) production at \(\sqrt{s} =\) 250 GeV (left panel) and \(e^+ e^-\rightarrow t \bar{t}\) production at \(\sqrt{s}=\) 3 TeV (right panel). The continuous line corresponds to the median relative correction, the dashed line to the mean. Results are shown for three algorithms: the generalised \(e^+e^-\) algorithm, the longitudinally invariant \(k_t\) algorithm and the VLC algorithm with \(\beta = \) 1. The statistical uncertainties of the results are negligible and are not indicated

4.4 Non-perturbative corrections

The largest part of the jet energy correction due to the finite size is amenable to perturbative calculations. A small residual correction is related to the hadronisation and must be extracted from (or tuned to) data. The non-perturbative energy correction is estimated as the difference between the energy of the parton-level jet, clustering all partons before hadronisation, and the jet reconstructed from stable final-state particles. The difference in energy between the parton-level and particle-level jet is typically small, but the distribution is offset from 0 and has a long asymmetric tail. Mean and median are again different and even have opposite signs.

The dependence of this correction on R is shown in Fig. 6. The non-perturbative part is very small compared to the total correction. It is well below 1% at \(\sqrt{s} = \) 250 GeV, for any value of R studied here. For high-energy collisions the correction is well below the per mille level. The generalised \(e^+e^-\) algorithm again has the best convergence, while for both VLC and longitudinally invariant \(k_t\) the median or mean remain sizeable even for \(R=\) 1.5.

4.5 Jet mass corrections

The previous discussion has focussed on the jet energy response. Corrections to other jet properties may also be important. Here, we study the corrections to the jet mass,Footnote 7 which can be taken as a proxy for the substructure of the jet. The non-perturbative jet mass correction is defined (analogously to the non-perturbative energy correction) as the difference between the masses of the parton-level and particle-level jet.

Fig. 8
figure 8

The response to 1.5 TeV top jets as a function of polar angle for three jet algorithms, all with radius parameter \(R=\) 1.5. The left plot shows the median reconstructed jet energy, the right plot the mean jet mass. Both quantities are normalised to the response of the Durham algorithm: \(E=\) 1.5 TeV, \(m\sim \) 370 GeV

The dependence on the radius parameter R is shown in Fig. 7. The non-perturbative contribution to the jet mass is quite large. The correction can be several tens of % at low energy. The relative correction is much reduced at higher energy: for \(e^+ e^-\rightarrow q \bar{q}\) production at \(\sqrt{s}=\) 3 TeVthe relative non-perturbative jet mass corrections are a factor three smaller than for the same process at \(\sqrt{s}=\) 250 GeV. Unlike the energy corrections, the jet mass corrections depend rather strongly on the process. In \(t\bar{t}\) production at \(\sqrt{s}=\) 3 TeV, where the jet mass ranges from the top-quark mass to several 100 GeV, the non-perturbative correction amount to a few %. In this case, the algorithms with the \(e^+ e^-\) inter-particle distance (generalised Durham and VLC) converge slightly faster than the longitudinally invariant algorithms.

5 Particle-level results

In this section the response of several algorithms is studied on simulated \(e^+ e^-\rightarrow t \bar{t}\) events at \(\sqrt{s} = \) 3 TeV. Clustering is exclusive, with \(N=\) 2. Both highly boosted top quarks are reconstructed as a single, large-R jet. We gain insight in the impact of the background by superposing randomly distributed background on the signal events. To this end the Monte Carlo setup described in Sect. 4.1 is extended with a simple mechanism to superpose a random energy flow on the signal event.

5.1 Jet energy response without background

Before we study the impact of the background, the response of the jet algorithms to the signal event is estimated. The energy response is determined as the median reconstructed energy. The mass distribution has a sharp peak at the top-quark mass and a long tail towards larger masses. The mass response is therefore estimated as the mean reconstructed jet mass. In both cases the response of the Durham algorithm – which clusters all final-state particles into the jets – is taken as a reference. The reconstructed energy is divided by 1.5 TeV, the reconstructed jet mass by the average jet mass of \(\sim \) 370 GeV.

In Fig. 8 the energy and mass response is shown as a function of polar angle for three algorithms: the generalised \(e^+e^-\) \(k_t\) algorithm (black), the longitudinally invariant \(k_t\) algorithm (blue dashed) and VLC with \(\beta =\gamma =\) 1 (red). The R-parameter is set to 1.5 for all three algorithms. The generalised \(e^+e^-\) \(k_t\) algorithm recovers over 99.9% of the top-quark energy for \(R=1.5\), independent of the jet polar angle. The shrinking jet areas in the forward region of the longitudinally invariant \(k_t\) and VLC algorithms lead to a slightly smaller response for \(|\cos {\theta }| > 0.6\). The polar angle dependence of longitudinally invariant \(k_t\) is more pronounced.

The mass response of all three algorithms is substantially lower than for the Durham algorithm. The generalised \(e^+e^-\) \(k_t\) algorithm has a flat response at nearly 80%. The VLC and longitudinally invariant \(k_t\) algorithms display the same pattern as for the energy response: VLC starts off with a lower response in the central region, but the response is much flatter versus polar angle.

5.2 Jet energy response with background

To gain insight in the performance in a more realistic environment with background, we overlay 200 1-GeV particles on each signal event. The background distribution is strongly peaked in the forward direction following an exponential distribution peaked at \( \theta = 0\), an approximation to the \(\gamma \gamma \rightarrow \) hadrons background in energy-frontier electron–positron colliders (a more realistic simulation of this background follows in Sect. 6.1).

Fig. 9
figure 9

Event display for a \(e^+e^- \rightarrow t\bar{t}\) event at \(\sqrt{s}=\) 3 TeV. The left panel shows the result of clustering with the longitudinally invariant \(k_t\) algorithm with \(R= 1.2\), the right panel the corresponding VLC jet with the same radius parameter. The image zooms in on the \(\theta - \phi \) area around one of the top jets. The location of the jet axis is indicated as a red circle. The area where the distance to the jet axis is smaller than the radius parameter (\(\Delta R_{iC}\) for longitudinally invariant \(k_t\), \( 1 - \cos \theta _{iC} \) for VLC) is indicated by the shaded region. The green squares represent particles from the top decay that are associated with each jet, the blue squares to background particles clustered into the jet

The two event displays in Fig. 9 provide a zoom image of the \(\theta - \phi \) plane for a single event. The location of the jet axis is indicated as a red circle. The approximate catchment area of both jets is shown in grey. The green squares represent particles from the top decay that are associated with each jet, the blue squares to background particles clustered into the jet. Both algorithms find a very similar jet axis, centered on the high-energy core of the jet. However, the algorithms have quite distinctive footprints. The longitudinally invariant algorithms expose a larger area in the forward region, which renders it more vulnerable to background in this region.

Fig. 10
figure 10

The average contribution to the jet energy (left plot) and jet mass (right plot) of 200 GeV of forward-peaked background

A quantitative view is obtained by comparing the energy and mass of jets obtained when clustering the same events with and without background particles. The bias (the average difference) in the jet energy and jet mass is shown in Fig. 10. The background leads to a significant bias for forward jets reconstructed with the longitudinally invariant \(k_t\) algorithm. The VLC algorithm, on the other hand, is only affected in the very forward region and the bias is much less pronounced.

The jet mass is known to be quite sensitive to soft and diffuse radiation, with the contribution scaling as the third power of the jet area [41]. We indeed find that the mass is strongly affected. A comparison of the jet reconstruction performance of the same process in a fully realistic environment is presented in Sect. 6.3.

6 Results from full simulation

The performance of the different algorithms is compared in full-simulation samples. We choose two benchmark scenarios with fully hadronic final states that challenge jet reconstruction: di-Higgs boson production (with \(h \rightarrow b\bar{b}\), i.e. a final state with four b-jets) and \( t \bar{t}\) production. Both analyses are performed at \(\sqrt{s}= \) 3 TeV, with the CLIC_ILD detector and a realistic background of \(\gamma \gamma \rightarrow \)hadrons.

6.1 Monte Carlo setup

The studies in this section are performed on CLIC 3 TeV Monte Carlo samples. Events are generated with WHIZARD [42] (version 1.95). The response of the CLIC_ILD detector [8] is simulated with GEANT4 [43]. Multi-peripheral \(\gamma \gamma \rightarrow \)hadrons events are generated with Pythia and superposed as pile-up on the signal events.

Fig. 11
figure 11

The reconstructed di-jet mass distribution for fully hadronic decays of \(e^+ e^- \rightarrow \nu \bar{\nu } hh\), \(h\rightarrow b\bar{b}\) events at a 3 TeV CLIC. In the left panel all Higgs boson candidates are included, in the right panel only those that match onto exactly two b-quarks from Higgs boson decay. The nominal level of \(\gamma \gamma \rightarrow \) hadrons background is overlaid on the signal. Particle-flow objects are selected using the tight selection

Table 2 Response and resolution of the di-jet mass distributions obtained with the \(k_t\) and VLC algorithms in the left panel of Fig. 11. The columns list the median di-jet mass, the \(\mathrm {IQR}_{34}\) and the \(\mathrm {RMS}_{90}\)

At CLIC bunches are spaced by 0.5 ns and detector systems are expected to integrate the background of a number of subsequent bunch crossings. In this study, the background corresponding to 60 bunch crossings is overlaid. In the event reconstruction, the information of the tracking system and the calorimeters is combined to form particle-flow objects (PFO) with the Pandora [44] algorithm. Timing cuts on PFOs reduce the background level, with a very small impact on the signal energy flow. The nominal (or default) selection of Ref. [3, 14] reduces the 19 TeV of energy deposited in the calorimeters by the entire bunch train to approximately 200 GeV superposed on a reconstructed events. A more stringent set of cuts, referred to as the tight selection in Ref. [3], reduces the background energy by another factor of two. Both scenarios are studied in the following.

The event simulation and reconstruction of the large data samples used in this study was performed using the ILCDIRAC [45, 46] grid production tools.

6.2 Higgs pair production

The study of Higgs boson pair production is crucial to assess the strength of the Higgs self-coupling. The analysis is very challenging at both hadron and lepton colliders due to the very small cross section. At an \(e^+ e^-\) collider, the significance of this signal is enhanced at large centre-of-mass energy, as the production rate in the vector-boson-fusion channel \(e^+ e^- \rightarrow \nu \bar{\nu } hh\) grows strongly with centre-of-mass energy. In this section we focus on events where both Higgs bosons decay to hadrons, through the dominant \(h \rightarrow b \bar{b}\) decay of the SM Higgs boson. The Higgs boson in the simulation has a mass of 126 GeVand Standard Model couplings.

This final state can be isolated [47] provided the four jets are reconstructed with excellent energy resolution. The challenge of this measurement lies in the fact that both Higgs bosons are typically emitted at small polar angle [28]. The most frequently observed topology has both Higgs bosons emitted in opposite directions: one in the forward direction and the other in the backward direction. At 3 TeV (at least) one of the Higgs bosons is emitted with \(|\cos {\theta }|>\) 0.9 in approximately 85% of events. In this area of the detector the background level due to \(\gamma \gamma \rightarrow \) hadrons production is most prominent.

Despite the large centre-of-mass energy, the Higgs bosons are produced with rather moderate energy: in 3 TeV collisions the most probable energy of the Higgs bosons is approximately 200 GeV, with a long, scarcely populated tail extending to 1.5 TeV. The modest Higgs boost is sufficient for the b-quarks to continue in the same hemisphere as their parent Higgs boson, but it is rarely large enough for the Higgs boson to form a single jet.

We perform exclusive jet reconstruction with \(N_{jets}=\) 4. The analysis is repeated for eight choices of the R parameter between 0.5 and 1.5. Higgs boson candidates are reconstructed by pairing two out of the four jets. The combination is retained that yields the best di-jet masses (i.e. that minimises \(\chi ^2 = (m_{ij} - m_{h})^2 + (m_{kl} - m_{h})^2\), where \(m_{ij}\) and \(m_{kl}\) are the masses of the two di-jet systems and \(m_h = \) 126 GeV is the nominal Higgs boson mass used in the simulation).

The distribution of the reconstructed mass of both di-jet systems forming the Higgs boson candidates is shown in Fig. 11. The results of two algorithms are shown, both with the radius parameter R set to 1.3. The red line denotes the result of the VLC algorithm with \(\beta = \gamma =\) 1, the blue line that of the longitudinally invariant \(k_t\) algorithm. Numerical results of the centre and width of the reconstructed di-jet mass distribution are presented in Table 2. The response of both algorithms is found to agree to within 0.5%, for all methods to estimate the central value of the distribution. The Higgs mass resolution obtained with the VLC algorithm is better for both figures of merit. The \(\mathrm {IQR}_{34}\) divided by the median yields 22.6% for the VLC algorithm versus 25.4% for \(k_t\).

Fig. 12
figure 12

The reconstructed di-jet mass resolution (determined as the 34% inter-quantile range \(\mathrm {IQR}_{34}\)) for simulated, fully hadronic decays of \(e^+ e^- \rightarrow \nu \bar{\nu } hh\), \(h\rightarrow b\bar{b}\) events produced in 3 TeV \(e^+ e^-\) collisions at CLIC. The nominal \(\gamma \gamma \rightarrow \) hadrons background is overlaid on the signal event. Particle-flow objects are selected using the tight selection

Fig. 13
figure 13

The jet energy residuals (reconstructed minus true energy) for fully hadronic decays of \(t\bar{t} \) events at a 3 TeV CLIC. No backgrounds are added in the left plot. In the right plot 60 bunch crossings of \(\gamma \gamma \rightarrow \) hadrons background are overlaid on the signal and particle-flow objects are selected using the tight selection

The dependence of the \(\mathrm {IQR}_{34}\) resolution on the radius parameter and the parameter \(\gamma \) of the VLC algorithm is shown in Fig. 12. The best mass resolution is obtained for large values of R in both algorithms. The choice of \(R \sim 1.3\) is close to optimal for both algorithms. Variation of the \( \gamma \) parameter, which controls the evolution of the VLC jet area in the forward region, leads to a shift of the optimal value of R. With \(\gamma < 1\) the jet area is reduced at a slightly slower rate and the best resolution is obtained for smaller R. Choosing \(\gamma > 1\) the jet area shrinks more rapidly and a larger R is required to capture the complete energy flow.

Table 3 The bias and resolution of the energy and mass measurements of reconstructed top jets in top-quark pair production with fully hadronic top-quark decay at a centre-of-mass energy of 3 TeV. Results are presented for the median response and two estimates of the resolution: the 34% inter-quantile range (\(\mathrm {IQR}_{34}\)) and the RMS of central 90% of jets (\(\mathrm {RMS}_{90}\)). All results are obtained by comparing the jet energy reconstructed from particle-flow objects to the jet of stable MC particles from the signal event. The performance of the classical \(e^+ e^-\) algorithm is such that the figures-of-merit cannot be estimated reliably under nominal background conditions (indicated by “–” entries in the table)
Fig. 14
figure 14

The reconstructed jet mass distribution for fully hadronic decays of \(t\bar{t} \) events at a 3 TeV CLIC. No backgrounds are added in the left plot. In the right plot 60 bunch crossings of \(\gamma \gamma \rightarrow \) hadrons background are overlaid on the signal and particle-flow objects are selected using the tight selection

6.3 Top quark pair production

The second benchmark we analyse is pair production of boosted top quarks in multi-TeV operation of the CLIC \(e^+e^-\) collider. At these energies the top-quark decay products are so collimated that hadronic top quarks can be reconstructed as a single large-R top-jet (\(R\sim 1\)). Only the fully hadronic final state \(e^+ e^- \rightarrow t \bar{t} \rightarrow b\bar{b}q\bar{q}'q''\bar{q}'''\) is considered. Events where either the top or anti-top quark is emitted in the forward or backward direction(\(|\cos {\theta }| > 0.7\)) are discarded to avoid the incomplete acceptance in that region. To cope with the increased background at 3 TeV the tight PFO selection is applied. Jets are reconstructed with exclusive (\(N=2\)) clustering with \(R=1.2\), which yields an adequate reconstruction of both the jet energy and the jet mass. For comparison the same algorithm is also run on all stable Monte Carlo particles. These include neutrinos, but not the particles from \(\gamma \gamma \rightarrow \) hadrons background.

The jet energy is a fairly good measure of the top-quark energy. The correction to the top-quark energy is typically 3.5% and the energy resolution is typically 8%. To measure the performance we compare the jet reconstructed from particle-flow objects with the jet found by the same algorithm on the stable particles from the signal event (i.e. excluding the \(\gamma \gamma \rightarrow \) hadrons). The jet energy residual is defined as the difference of the energy of detector-level and particle-level jets. The distribution is shown in Fig. 13. The response is measured as the median of the residual distribution. The resolution is measured as the \(\mathrm {IQR}_{34}\).

Quantitative results are presented in Table 3. The \(\mathrm {RMS}_{90}\) is also presented to facilitate comparison to other studies. In the absence of background, all algorithms reconstruct the energy of the jet quite precisely, with a bias of less than 1% and a resolution of 2–4%. The performance of the classical \(e^+ e^-\) algorithms is degraded as soon as the \(\gamma \gamma \rightarrow \) hadrons background with tight PFO selection is added. The VLC and longitudinally invariant \(k_t\) algorithms show very little performance degradation even with the nominal PFO selection.

The jet invariant mass is much more sensitive to soft background contamination [39, 41]. The jet mass distributions are presented in Fig. 14. In the left panel, which corresponds to \(t\bar{t}\) events without background overlay, all algorithms are seen to reconstruct a narrow peak close to the top-quark mass. The long tail toward large mass is due to radiation off the top quark and its decay products and is also present in the jets reconstructed from stable MC particles. The plots in the right panel show a severe degradation when the \(\gamma \gamma \rightarrow \) hadrons background with tight PFO selection is added, most noticeably for the Durham algorithm. The bias and resolution of the jet mass is shown as a function of radius parameter in Fig. 15.

A quantitative summary is presented in the second part of Table 3. The bias on the jet mass without background is sub-% for most algorithms. The resolution of the VLC and longitudinally invariant \(k_t\) algorithms is significantly better than that of the classical \(e^+ e^-\) algorithms. The 4.1% resolution is a testimony to the potential of highly granular calorimeters and particle-flow reconstruction for jet substructure measurements.

The \(\gamma \gamma \rightarrow \) hadrons background has a profound effect on the performance. The performance of the classical algorithms is clearly inadequate, with a strong bias and a severe degradation, even with the tight PFO selection. The VLC and longitudinally invariant \(k_t\) algorithms are much less affected, as expected from the smaller exposed area. The VLC algorithm is found to be more resilient than the longitudinally invariant \(k_t\), confirming the result anticipated at the particle level in Sect. 5.

Fig. 15
figure 15

The bias and resolution of the reconstructed jet mass versus radius parameter R of two jet algorithms. The jets are reconstructed in fully hadronic top-quark decays in \(t\bar{t}\) events at a center-of-mass energy of 3 TeV. The jets reconstructed from particle-flow objects are compared to jets reconstructed from all stable particles from the signal event. The black curves correspond to the results obtained without background, the blue dashed and red curves to 60 bunch crossings of \(\gamma \gamma \rightarrow \)hadrons background overlaid on the signal, with the tight PFO selection

7 Discussion

In this section we discuss the implications for lower-energy colliders and identify several topics that merit further study.

7.1 Implications for lower-energy lepton colliders

In this paper we have focussed on CLIC operation at \(\sqrt{s} = \) 3 TeV, arguably the most challenging environment that lepton colliders might face in the next decades. We have chosen this environment because subtle differences in jet definitions lead to significant differences in performance. This has helped us to gain a deeper understanding of the intricacies of jet reconstruction at lepton colliders and to establish solid conclusions about the resilience of the different algorithms.

These findings are by no means limited to the CLIC environment. Subtle but significant differences in performance are expected also for the ILC at 250 or 500 GeV and at circular colliders.

7.2 Further R&D on jet algorithms

The set of jet algorithms studied in this paper is by no means exhaustive. This study does not address several recent proposals, such as the XCone algorithm [48] or the global jet clustering proposed by Georgi [49].

A broad range of new techniques developed for the LHC have so far remained unexplored. This is particularly true for a set of tools that has proven extremely powerful in pile-up mitigation and correction in ATLAS and CMS.

Jet grooming (the collective name for (mass-drop) filtering [50], pruning [51] and trimming [52]) effectively reduces the exposed jet area to several small regions with large energy flow. This provides an effective means of capturing a large fraction of the jet energy while reducing the impact of soft contamination. Tests of the trimming algorithm in the CLIC environment yield very good results, improving the jet mass resolution significantly.

Techniques to correct for the effect of pile-up based on an event-by-event measurement of the pile-up activity [53] are quite successful at the LHC. Subtraction at the constituent level with dynamical thresholds [54,55,56] is under active development. An adaptation to the environment at lepton colliders, with a very sparse background energy flow, may prove useful.

8 Conclusions

We have studied the jet reconstruction performance of several sequential reconstruction algorithms at high-energy lepton colliders. In addition to the classical \(e^+ e^-\) algorithms we include a version of the same algorithms with beam jets. We also study the performance of the longitudinally invariant \(k_t\) algorithm and a new \(e^+ e^-\) algorithm, called VLC [21], which are expected to be more resilient to the impact of backgrounds.

The study is based on detailed Monte Carlo simulation. For two benchmark processes we use a full simulation of the linear collider detector concepts, including the relevant background processes.

The perturbative energy corrections of all algorithms with finite size jets are sizeable for small values of the radius parameter (10–15% for \(R=0.5\)) and decrease to 1–5% for \(R= 1.5\). This result is approximately independent of the process and centre-of-mass energy. Convergence with R is faster for the generalised \(e^+ e^-\) algorithm than for the longitudinally invariant \(k_t\) algorithm and VLC, which expose a smaller area to forward jets.

The non-perturbative hadronisation correction represents a very small part of the total correction. Its relevance decreases with increasing centre-of-mass energy: it is less than 1% at \(\sqrt{s}=\) 250 GeV for all algorithms and less than a per mille at \(\sqrt{s}=\) 3 TeV. The generalised \(e^+ e^-\) algorithm again converges fastest. We have estimated the non-perturbative correction also for the jet invariant mass: these corrections are much larger than the non-perturbative energy corrections and remain of the order of a few % even for \(R= 1.5\) and \(\sqrt{s}= 3\) TeV.

The forward-peaked \(\gamma \gamma \rightarrow \)hadrons background at future high-energy linear lepton colliders is one of the most important factors in the jet reconstruction performance. It has motivated ILC and CLIC to abandon classical, inclusive algorithms in favour of algorithms with a finite jet size. Algorithms that expose a reduced solid angle in the forward region of the detector, such as longitudinally invariant \(k_t\) or VLC are more robust. A particle-level study shows that these two algorithms have a different response, with VLC showing a lower response, but one that is more stable versus polar angle. VLC is found to be less susceptible to background.

We present two studies in full simulation, namely di-Higgs production and top-quark pair production at \(\sqrt{s} =\) 3 TeV, which present a combination of a relatively harsh background level, high jet multiplicity and forward jets. In both cases the classical \(e^+ e^-\) algorithms offer an inadequate performance. The same is true for the generalised version with beam jets. VLC provides significantly better mass resolution for the Higgs study and considerably better jet mass reconstruction than the longitudinally invariant \(k_t\) algorithm.

Jet clustering is key technique for many analyses of multi-jet final states at future high-energy electron–positron colliders. This study shows that a considerable increase in performance can be obtained by a careful choice of the clustering algorithm. We recommend, therefore, that studies into the physics potential of future \(e^+ e^-\) colliders carefully optimise the choice of the jet reconstruction algorithm and its parameters. We also encourage further work on robust algorithms for \(e^+ e^-\) collisions.