Buckets of Tops

Reconstructing hadronically decaying top quarks is a key challenge at the LHC, affecting a long list of Higgs analyses and new physics searches. We propose a new method of collecting jets in buckets, corresponding to top quarks and initial state radiation. This method is particularly well suited for moderate transverse momenta of the top quark, closing the gap between top taggers and traditional top reconstruction. Applying it to searches for supersymmetric top squarks we illustrate the power of buckets.


I. INTRODUCTION
An important difference between the Tevatron and LHC is that the latter can produce and study top quarks in great numbers [1]. This allows us to investigate all different top production mechanisms in detail, including their QCD structure. After the discovery of a Higgs-like [2] resonance, studying its coupling to the top quark will play a particularly critical role in our understanding of the Higgs sector. This is made most obvious in the renormalization group evolution of the Higgs potential to large energy scales [3]. The direct measurement of the top Yukawa coupling clearly hinges on top quark identification and reconstruction. At the same time, we have reason to suspect that new physics that solves the hierarchy problem and lives at sufficiently high energy scales tends to couple strongly to top quarks [4]. This motivates us to search for new physics in the LHC top sample; for example by searching for top pair resonance structures of top pair production associated with missing transverse momentum.
Historically, the study of top pair production has largely been restricted to semi-leptonic decays of the two tops quarks. The reason is that the lepton effectively removes the overwhelming QCD background. However, purely leptonic top pairs not only come at a much smaller rate, they also include two neutrinos, challenging any analysis based on the observed missing transverse momentum. A major challenge in top physics at the LHC is how to gain access to the purely hadronic decays of top quarks.
Identifying hadronic top decays using a jet algorithm was part of the original proposal of jet substructure analyses [5,6]. Some of the early jet substructure algorithms were designed to target hadronic top decays [7]. While Higgs taggers [8] should clearly have a high priority within the LHC experiments, working top taggers are the perfect laboratory to test how well substructure approaches work in practice.
The moment we go beyond searches for heavy resonances the main problem of all top taggers is the size of the initial fat jet. For example, using the Cambridge-Aachen jet of size R = 1.5 as the starting point of the HEPTopTagger [9][10][11] limits the momentum range of reconstructable top quarks to p T,t 200 GeV. Essentially, all other top tagging approaches require even higher boost. Increasing the size of the fat jet to R = 1.8 raises several QCD and combinatorics issues [12]. The big question in using hadronic top analyses as part of Higgs searches or top partner searches is how to further reduce this top momentum threshold.
In this paper we propose an alternative method for an efficient top reconstruction at moderate momentum. It targets the transverse momentum regime, in the fully hadronic decay mode. Starting from an event with a high multiplicity of jets, we assign all jets into three groups or 'buckets'. The buckets are chosen based on a metric in terms of invariant masses, defining two top buckets and a third bucket containing the extra hadronic activity like initial state radiation (ISR). While initially this search strategy does not prefer boosted top quarks, we will see how such events are eventually preferred from a combinatorics perspective.
In Section II we start with a simple algorithm for reconstructing tops in buckets. We test this algorithm for hadronically decaying top pairs as well as W +jets and pure QCD jets backgrounds. Additional handles will help us separate the top signal from the backgrounds. In Section III, we modify the simple algorithm to take advantage of the b quarks and W bosons that are present in top decays but not in the QCD backgrounds. This improved bucket algorithm is optimized to efficiently find and reconstruct top pairs with moderate p T . In Section IV we apply our bucket algorithm to stop pair searches.

II. SIMPLE BUCKET ALGORITHM
In this section, we start with a simple algorithm to identify and reconstruct hadronically decaying top pairs. While an improved algorithm will be presented in the next section, this simple version captures many of the key concepts we will employ later. The overall scheme is fairly straightforward: by assumption every jet originates from one of the two tops or from initial state radiation, so we assign every jet to one of three 'buckets'. Jets in buckets B 1 and B 2 correspond to top decays, while all remaining jets are placed in B ISR . We cycle through every permutation of jet assignments to minimize the distance between the invariant masses of the jets in B 1 and B 2 and the top mass. The metric is chosen to ensure that bucket B 1 reconstructs the top mass better than bucket B 2 .
Here and throughout the remainder of the paper, all Standard Model (SM) samples are generated with Alpgen+Pythia [13,14]. We use matrix-level matching [15] to correctly describe jet radiation over the entire phase space. This includes up to tt+2 jets, W +4 jets and 3−5 QCD jets, with the top cross sections normalized to next-to-next-to-leading order [16]. Jets are reconstructed using the Cambridge/Aachen algorithm [17] of size R = 0.5 in FastJet [18]. Note that all our results are relatively insensitive to the choice of jet algorithm.
All leptons we require to be hard and isolated: p T, > 10 GeV and no track of another charged particle within R < 0.5 around the lepton. We consider only jets with p T > 25 GeV and |η| < 2.5. Even though the algorithm presented in this section is in principle applicable to events with any number of jets we preselect events with five or more jets to reduce QCD backgrounds. Because we are interested in hadronically decaying tt pairs we veto on isolated leptons. The restricted sample denoted as t hth has a cross section of 104 pb at the LHC with √ s = 8 TeV. One last word concerning underlying event and pile-up: unlike methods involving jet substructure [6] our bucket reconstruction relies on standard jets with moderately large multiplicities, so aside from jet energy scale uncertainties we do not expect specific experimental or theoretical challenges.

Bucket definition
As the goal of the bucket algorithm is to identify tops by sorting jets into categories that resemble tops, we need a metric to determine the similarity of a collection of jets to a top. For simple buckets B i it is where we sum over all four-vectors in the bucket. For each event with five or more jets we permute over all possible groupings of the jets into three buckets {B 1 , B 2 , B ISR }. We then select the combination that minimizes a global metric defined as The factor ω > 1 stabilizes the grouping of jets into buckets. In this work we take ω = 100, effectively decoupling ∆ B2 from the metric. As a consequence we always find ∆ B1 < ∆ B2 , i.e. B 1 is the bucket with an invariant mass closer to that of the top than the invariant mass of bucket B 2 . Other values of ω might eventually turn out more appropriate for different applications.
As the first selection cut we require the invariant masses of both top buckets, B 1 and B 2 , to lie in the window The lower limit selects events above the Jacobian peak for top decays. We will see that this selection improves the top signal over QCD background S/B by about a factor of two. All buckets passing Eq. (4) we categorize by their number of jets; buckets including three or more jets (3j-buckets) and those including two jets (2j-buckets). Selecting only events with two 3j-buckets improves the signal-to-background ratio by a factor of five.

Jet selection
For tagging two tops in fully hadronic mode, we might naively require at least six reconstructed jets. In practice, with a threshold of p T,j > 25 GeV this condition is too strict. To improve our efficiency we need to consider the case where one of the jets from top pair decays is missing. It is also worth noting that even requiring six jets does not guarantee that we collect all six decay products of the top pair. Frequently, some of the observed jets come from initial state radiation instead [12].
In Figure 1 we plot the parton level p T distributions of the six decay partons from the top pairs. In the left panel we see that the four hardest decay jets are not affected by the threshold p T,j > 25 GeV. In contrast, the softest distribution only peaks around 25 GeV, so roughly half the events do not pass our threshold on the sixth jet. Table I shows the number of events in the hadronic t hth sample after several cuts on the jet multiplicity, and the percentage of events with the 5 th or 6 th parton-level top decay jets above p T,j > 25 GeV. In about a half of events with at least six jets the sixth top-decay parton falls below the p T threshold. Adding the two columns tells us that more than 90% of all events capture five of the six top decay products. Requiring only five instead of six jets increases the fraction of events where we miss only one of the top decay products to almost half. The  top p T threshold our central values for the efficiencies are not strongly affected, but hadronization as well as detector effects might lead to significant shifts due to the steep p T,j behavior.

W reconstruction
After placing each of the jets in the event into one of three buckets (B 1 , B 2 , or B ISR ) we require the 3jbuckets to contain a hadronically decaying W candidate. In the rare case of one bucket consisting of more than three jets we combine them into exactly three jets using the C/A algorithm and then look for a W candidate. As in the HEPTopTagger [10] we define a mass ratio cut for at least one combination of jets k, in the bucket i. Events with 2j-buckets by construction cannot satisfy Eq. (5). In addition, in such events one of the W decay jets is typically the softest jet and does not pass the p T threshold, and so the W reconstruction could not occur regardless In our first, naive approach we categorize all events with two valid top buckets into three types: • (t w ,t w ): both top buckets have W candidates as defined by Eq. (5), : only the first or second top bucket has a W candidate, The t w or t − status is ordered as (B 1 , B 2 ), where B 1 is defined as the bucket closest in mass to the top. Buckets classified as t w have to be 3j-buckets, while t − buckets can be either 3j or 2j.
To extract hadronic top pair events from the QCD background we can compare the different categories on Monte-Carlo truth level. Starting from S/B ∼ 0.005 after the lepton veto and selecting only (t w ,t w ) events yields the highest value S/B ∼ 0.09. This corresponds to an improvement of S/B by almost a factor 20. To improve beyond this level, we need to require at least one, preferably two, b-tags to control the mostly Yang-Mills and light-flavor QCD background.

b-tags
To further reduce the QCD background we exploit b-tags. We assume b-tagging and mis-tagging efficiencies for light flavors ( b , mis ) to be (70%, 1%), and fully account for combinatorial factors in the background. For the tt+jets signal the effect of mis-tagging is sub-leading and can be ignored.
To avoid combinatorics we could impose b-tagging only for the most likely b-jet in a bucket based on the W condition factors to improve S/B, as suggested in Ref. [12]. In this algorithm we do not take this option because it reduces the signal efficiency. We prefer to keep the maximum fraction of signal events especially for the case that both signal and the main background include tt events, such as the top partner searches discussed below.
In any top-tagging algorithm we are interested not only in extracting the signal from backgrounds, but in accurately reconstructing the original top momenta. For a measure of our reconstruction accuracy we use the geometric distance in the (η, φ) plane between the bucket momentum and the closer top parton momentum p t obtained from Monte-Carlo truth, We consider our reconstruction successful when R i < 0.5. In the following, we indicate the percentage of events with both buckets reconstructing top momenta (R 1 < 0.5 and R 2 < 0.5), events where only B 1 reconstructs the top momentum (R 1 < 0.5 < R 2 ), and events where only B 2 reconstructs the top momentum (R 2 < 0.5 < R 1 ). The last case allows for events where the second bucket (with its worse top mass reconstruction) actually gives the better top direction.  For (t w ,t w ) events where each bucket contains exactly one b-jet, the top momentum reconstruction is generally good. As seen in to Table II about 75% of the events reconstruct both top directions well. As expected from the discussion above, a significant fraction of signal events only give (3j,2j)-buckets. When a W candidate is not found, but each bucket contains a b-tag and lies in the top mass window Eq. (4), the momentum reconstruction is good only for the t w bucket; in these events half of the t w momenta with a W candidate reconstruct the top direction well. All this points to using the b-tag information to improve our reconstruction algorithm. This will be the starting point of the improved algorithm in the next section.

III. BOTTOM-CENTERED BUCKETS
In Section II we have seen that we need at least two b-tags per event to control the QCD background. However, in the simple algorithm, each bucket does not always have exactly one b-jet and the reconstruction is not particularly effective for (t w ,t w ) events. The obvious solution is to define buckets around b-tagged jets, i.e. starting each bucket with the bottom jets (which are usually the hardest jets in the event) and adding light-flavor jets to it.
In this section we define buckets starting with the requirement that B 1 and B 2 each have exactly one b-jet, and restrict the possible permutation of jet assignments to B 1 , B 2 , and B ISR accordingly. Other than this, we use the same distance measure defined in Eq. (2) and Eq. (3) and select {B 1 , B 2 , B ISR } giving the minimum ∆. Figure 2 shows the bucket masses m B1 , m B2 and m ISR . For both tt and QCD samples the m B1 distributions peak at m t by construction. The distribution is narrower for the signal. The dip in the m B2 distributions at m t is due to the large weighting factor ω in Eq. (3), which defines the bucket with mass closest to m t to be B 1 . Compared with the m B1 distributions, the m B2 distributions are broad but still tend to peak toward m t .
As mentioned above, the analysis of the top buckets constructed around the b-tagged jets is the same as the    simple algorithm described in Section II, including the bucket mass cut in in Eq. (4). In Table III we show the corresponding results. Starting with two b-jets improves the number of (t w ,t w ) events by almost 50%. Roughly 70% of (t w , t w )-events reconstruct both tops well, essentially unchanged from the earlier analysis. One kind of events which is now correctly accounted for are cases where the simple algorithm finds two b-jets in the same bucket, and give a bucket mass in the correct range.
Asking for two b-tags within at least five jets at the very beginning produces large combinatorial factors for mis-tagging QCD background events. As a result the backgrounds double in each category and S/B degrades for (t w ,t w ) events.
While there is no obvious way to improve the (t w ,t w ) category of events, Table III shows that a significant number of events come out (t w ,t − ) and (t − ,t w ), that is, only one bucket contains a W candidate. For these events, the QCD background is not huge, S/B ∼ 3, so we will try to improve our treatment of this fraction of events.

b/jet Buckets
In Section II we found that it is not rare for the softest top decay jet to fall below the jet p T threshold. Attempts to reconstruct two tops in (3j,3j)-buckets will then fail. In 94% of these cases the softest of the six top decay partons comes from the W decay. Restricted to events where the sixth parton falls below 25 GeV this fraction increases to 98.5%, i.e. whenever the sixth parton is missing the surviving two jets are the bottom and the harder W decay jet. In Figure 3 we first show the invariant mass of the b and the harder W decay product m bj1 at parton level. We see a clear peak and an endpoint m bj1 < m 2 t − m 2 W 155 GeV [19]. For events where the softer W decay jet falls below the p T threshold the peak becomes more pronounced.
The question is: can we use the predicted peak in the m bj distribution to identify tops in 2j-buckets? If the third missing top decay jet indeed fails the p T threshold we expect the top momentum to be close to the b/jet momentum. The left panels of Figure 4 show the difference between the parton-level top momentum and the b/jet system in terms of (p T,bj − p T,t )/p T,bj ≡ ∆p T /p T,bj and ∆R. If we assume an R separation of around 0.5 as a quality measure the result looks promising. Similarly, we should be able to reconstruct the transverse momentum of the top at least at the 20% level without including the softest W decay jet. In comparing our bucket method to top taggers, it should be emphasized that for the bucket method we allow for a missing W decay jet rather than replacing the lighter W decay jet by a QCD jet [12].
For 2j-buckets, which we know do not include all three top decay products, we replace the distance measure of Eq. (2) with a similar measure inspired by the distribution of m bj , The peak value of 145 GeV is read off Figure 3 and should eventually be tuned to data. Because 3j-buckets already reconstruct the top momentum we keep them. For top buckets in the (t w ,t − ), (t − ,t w ), and (t − ,t − ) categories which do not contain a W candidate we re-assign jets replacing Eq. (2) with the new distance measure. In addition, we need to remove the top mass selection cut Eq. (4). This way combinations of b quarks and jets which do not fall into the window of Eq. (7) are kept. The new reconstruction algorithm reads • (t w ,t w ): keep these buckets as is, • (t w ,t − ) or (t − ,t w ): reconstruct the failed bucket using all non-t w jets, minimizing ∆ bj B ,  • (t − ,t − ): use all jets to minimize ∆ bj B1 + ∆ bj B2 . Note that for reconstructing b/jet-buckets we use jets both from the t − bucket and from the ISR bucket.
Comparing to the original algorithm we have adapted the metric for assigning jet for top buckets in the t − category. What remains is to replace the top mass window in Eq. (7) with appropriate b/jet values. In the right panel of Figure 3 we show the b/jet bucket mass distributions m bj for signal and background. For the signal they agree well with the expectation from the left panel of Figure 3. For a top candidate we require at least one b/jet pair satisfying 75 GeV < m bj < 155 GeV .
We show the signal and background efficiencies of this new reconstruction algorithm in Table IV, along with the percentage of correct top reconstruction. The numbers need to be compared to Table III. First, we see that the number of events which contain valid top buckets in the correct mass window, albeit including one 2j-bucket, has significantly increased. In the (t w ,t − ) category roughly half of all events reconstruct both tops well, in spite of missing one of the six decay jets. The number of (t − ,t w ) events passing this reconstruction algorithm drops significantly when compared to Table III. Most of these events contain one b-jet and one non-b-tagged jet in B 1 . However, the b-jet in this category of events is typically a merger of a b and the third jet from the top decay. Thus, while the bucket itself has an invariant mass near the top, it contains neither a W candidate nor a b-jet that can be combined with another jet in the event to pass the selection criteria in Eq. (8). Even in the (t − ,t − ) category where neither of the two buckets include a reconstructed W candidate the fraction of well reconstructed top pairs reaches almost 40%.
To study the quality of the top reconstruction in more detail we show the difference between the bucket momentum and the parton level top momentum in terms of ∆R and ∆p T /p T in the right two panels of Figure 4. The buckets constructed around b-jets are shown in black. The results of replacing the t − buckets using the b/jet algorithm are shown in red. In this case we see a narrow peak at zero which corresponds to complete top momentum reconstruction where we fail to find a W candidate due to overlapping jets. Such events -which are in the minority -often fail to pass the reconstruction using the ∆ bj B metric. As a result, the narrow peak at zero is not present in this second reconstruction method.
For t − buckets the b/jet algorithm consistently reconstructs the top direction significantly better than using the original method. In contrast, changing t w buckets to the b/jet-bucket does not improve the momentum reconstruction. We checked b/jet-momentum provides better top momentum reconstruction than only using the bottom momentum.

pT dependent efficiencies
Until now we have focused on identifying and reconstructing pairs of hadronically decaying top quarks from the complete signal sample. The results shown in Table IV indicate that the efficiency as well as the background rejection of our algorithm allows for a systematic study of hadronic top pairs. However, the fraction of events with not-quite-perfect reconstruction of the top directions (R i > 0.5 for i = 1, 2), is somewhat worrisome. From top tagging we know that a certain fraction of relatively poorly reconstructed tops cannot be avoided [12], but that fraction should be small. What we need is a self-consistency requirement -or QMM 1 -similar to only accepting reconstructed tops with p T,t > 200 GeV in a top tagger [10].
Once we identify a top buckets we can use two observables to define such a QMM: the top momentum and the geometric size of the hadronic top decay. The latter is defined differently for t w buckets and t − buckets. In the first case we have access to all pair-wise ∆R distances between the three top decay products. We define R bjj as the maximum of the three ∆R separations of the top decay products. For t − buckets we only have one distance, namely R bj between the bottom and the hardest light-flavor jet.
In Figure 5 we show the correlation between these two observables, first for parton level simulations in the left column. For both kinds of buckets we see a clear correlation, with the main difference being that most t w buckets have relatively low transverse momenta. For t − buckets, which require the softest top decay jet to fall below p T,j = 25 GeV, the distribution extends to larger transverse momenta where the initial boost of the top can compensate or the decay momentum of the softest jet.
The second column shows the reconstructed observables for t w and t − buckets, requiring that the buckets reconstruct the parton level top direction within R < 0.5. The correlation between size and transverse momentum is the same as expected from simulation. However, we clearly see that either large transverse momenta, p T,t 100 GeV, or small sizes, ∆R bj(j) 2.5, are preferred. This is particularly true for t − buckets. The reason for this is that a slight boost of the top quarks generates a geometric separation of the transverse backto-back tops and the forward ISR jets. Combinations of jets from different buckets are now separated in their typical transverse mass values. This gives us a handle on combinatorics and improves the top reconstruction even in the case where one of the top decay products is missing. corresponding efficiency for two HEPTopTagger [9,10] tags. In all cases the last bin includes all events above 450 GeV.
Conversely, buckets passing as tops but giving a poor directional reconstruction reside at low transverse momenta and large size, as can be seen in the third column of Figure 5. To veto these buckets we have a choice of criteria in the two-dimensional R bj(j) vs. p T,t plane. We choose the condition p rec T,t > 100 GeV (9) at the level of the buckets to increase the fraction of well reconstructed or matched top quarks in both bucket categories. This choice results in the highest efficiency of well-reconstructed tops in both t w and t − buckets. Alternative conditions in terms of R bj(j) or in the two-dimensional planes shown in Figure 5 could replace Eq. (9) in specific analyses. For example, a stricter cut will result in higher purity of well-reconstructed tops.
To illustrate the power of the bucket algorithm we compute the efficiency for reconstructing a single top as well as a top pair as a function of the transverse momenta of the tops. The left panel of Figure 6 shows the efficiency for a bucket tag as a function of the true transverse momentum of the top. The baseline is all fully hadronic tt events in the Standard Model, with five or more jets and two b-tags. A possible mis-measurement of p T,t in particular at low transverse momenta explains the tail of events below the apparent consistency criterion p rec T,t > 100 GeV. We see that the tagging efficiency increases rapidly right at threshold. Above p T,t = 150 GeV more than 90% of the tagged top quarks can be matched to a true top within R i < 0.5. For p T,t = 100−150 GeV about 80% can be so matched. For t w and t − buckets the number of unmatched tops becomes negligible above 250 GeV. Adding t w and t − buckets, the total efficiency of our algorithm is 60-70% for 150 < p T,t < 350 GeV.
In the central panel of Figure 6 we show the tagging efficiency for two top quarks as a function of the average true transverse momentump T = (p T,t1 + p T,t2 )/2. The total efficiency is split between (t w ,t w ) events (black), (t w ,t − ) or (t − ,t w ) events (red), and (t − ,t − ) events (green). For each of these categories we also show the well reconstructed tops only. As expected, the (t w ,t w ) events are reconstructed with an encouragingly high efficiency and essentially negligible number of non-matched tops. For the other two categories the fraction of unmatched tops is slightly larger, but well under control.
Also note that the efficiency for (t w ,t w ) events is slightly higher than the square of the single bucket t w efficiency. This is because, once one top in an event is reconstructed, the second top becomes easier to find, due to combinatorial factors. Similar correlations occur in the (t w ,t − ), (t − ,t w ) and (t − ,t − ) categories. The total double top tag efficiency forp T,t = 150 − 350 GeV is close to the single tag efficiency: 55-70%. As we always search for two tops (otherwise we regard the event as un-reconstructed), the total double tag efficiency and total bucket tag efficiency must be closely related, as long as the individual p T,t and averagedp T,t distributions are similar. We should note that some of the unmatched tops may still be correct tags as QCD effects will change the direction of the true top as compared to the top decay products at parton and particle level.
The resulting cross sections of reconstructed tops with the consistency selection cut Eq. (9)    in Table V. The total double top tag efficiency for the t hth +jets sample with five jets of which two are b-tagged is 28%. The mis-tagging efficiency of finding two valid top buckets in the pure QCD events (five jet, two mis-tagged as b-jets) is of the order of 5%.
Unlike for a typical top tagger, illustrated in the right panel, the efficiency of the buckets does not reach a plateau at large transverse momentum. Once the top decay jets start merging at the scale of the C/A jet size the method will fail, so for example R C/A = 0.5 leads to a drop above p T,t m t /R ∼ 350 GeV. Towards smaller top momenta the requirement Eq. (9) limits the efficiency by removing poorly reconstructed tops due to combinatorics. By construction, the bucket method targets the intermediate regime 150 GeV < p T,t < 350 GeV where it should serve as a very useful tool in Higgs searches as well as new physics searches.

IV. STOPS FROM BUCKETS
As a demonstration of our algorithm for top reconstruction, we apply it to scalar top searches. Searches for supersymmetry, or general, top partners are becoming more and more central in ATLAS and CMS. They constrain the allowed stop masses to mt 600 GeV [20]. Theoretically, many analysis strategies have been suggested, covering the semileptonic decay channel [21], the hadronic decay channel [22], or dedicated HEP-TopTagger studies in each of these channels [23].
In this section, we assume scalar top pair production followed by decay into tops and the lightest neutralinõ χ 0 1 with 100% branching ratio. For all model points we set the lightest neutralino mass to mχ0 1 = 100 GeV. Cross sections at the LHC assuming √ s = 8 TeV are shown in Table VI. To generate the signal for stop masses of 500, 600, and 700 GeV we use Herwig++ [24]. We normalize the production cross section to the Prospino results at next-to-leading order [25].
Since the reconstruction technique described in the previous section are also applicable for tops from stop decays, we expect good top reconstruction. To reduce the non-top background we first need to apply a set of simple selection cuts. We first require at least five jets, two of them b-tagged. Then, we require large missing momentum, / E T > 150 GeV, and veto isolated leptons. The results are summarized in Table VI. Because QCD has no intrinsic source of missing momentum, and W +jets has a small rate and a lepton we ignore this backgrounds in this paper, and assume mostly tt backgrounds with large missing transverse momentum, typically the result of mismeasurement or τ decay.
Based on the algorithm developed in this paper we require two top buckets with b/jet re-ordering. The two reconstructed bucket momenta we denote as p t1 and p t2 . After the missing momentum cut the main background is semi-leptonic top pairs, which means one of the two tagged tops in the background sample is mis-tagged.
The advantage of an analysis based on fully hadronic top decays is that both tops are fully reconstructable [10,11]. We use the bucket momenta to compute m T 2 (p t1 , p t2 , / E T ) [26]. Its distributions for the tt background and the stop pair signal is shown in Figure 7. To extract stop pairs we select events with After this cut and for a stop mass of 600 GeV we arrive at S/B ∼ 1 and more than three sigma significance at the 8 TeV LHC with the currently available integrated luminosity of 25 fb −1 . In addition, the endpoint of the m T 2 distribution with fully reconstructed hadronic tops should allow us to precisely measure the stop mass [10]. All intermediate steps as well as results for other stop masses are shown in Table VI. Note that some numbers are different from those shown in Table V due to the leptonic decays.
Of all events with two reconstructed tops about 10% involve τ leptons, both for the signal and the background. After the missing momentum cut a significant fraction (∼ 75%) of the top background comes from these events. In contrast, only 10% of the signal events include a top decay to a τ . Therefore, a τ -rejection would improve our results significantly, as shown in Table VI.

V. CONCLUSION
In this paper we have presented a new method to identify and reconstruct hadronically decaying top quarks. It is based on assigning regular jets to buckets, one for each top decay and one for initial state radiation. The buckets corresponding to tops are each seeded with one of the two b-jets we require in every events. If a top bucket includes all three top decay products it has to fulfill W and top mass constraints. However, frequently the softer W decay jet is missing, so we have to rely on the two leading jets to reconstruct a defined fraction of the top mass. After an appropriate re-ordering of the buckets missing the softest decay jets both kinds of buckets can be used to reconstruct the top four-momentum.
To suppress tops which for one or another reason cannot be matched to a generated top quark we apply a self consistency condition (QMM) to each bucket. This condition defines the lower bound of the typical transverse  momentum range 100 GeV < p T,t < 350 GeV to which the method is sensitive. For higher boosts the buckets will eventually fail due to the size of the jets they are constructed from. For top quarks with this moderate boost we achieve a maximum efficiency around 60-70% for the reconstruction of two top quarks. In particular, for p T,t < 250 GeV our method gives a significant improvement over subjet-based top taggers, which have low efficiencies in this regime.
To illustrate our approach in a new physics framework we have applied it to supersymmetric stop searches, relying on stop decays to tops and missing energy. Because we reconstruct the top four-momenta we can apply a simple m T 2 analysis, including a measurement of the stop mass. This makes stop search strategies as simple as sbottom or slepton searches.
While the detailed numerical results for our method should be tested in a realistic experimental environment there obviously exists a wide range of possible applications for top buckets in ATLAS and CMS. As a first step, hadronic top pair production with and without contributions from beyond the Standard Model might serve as a useful testing ground [11].