Abstract
We study the effectiveness of theoretically-motivated high-level jet observables in the extreme context of jets with a large number of hard sub-jets (up to N = 8). Previous studies indicate that high-level observables are powerful, interpretable tools to probe jet substructure for N ≤ 3 hard sub-jets, but that deep neural networks trained on low-level jet constituents match or slightly exceed their performance. We extend this work for up to N = 8 hard sub-jets, using deep particle-flow networks (PFNs) and Transformer based networks to estimate a loose upper bound on the classification performance. A fully-connected neural network operating on a standard set of high-level jet observables, 135 N-subjetiness observables and jet mass, reach classification accuracy of 86.90%, but fall short of the PFN and Transformer models, which reach classification accuracies of 89.19% and 91.27% respectively, suggesting that the constituent networks utilize information not captured by the set of high-level observables. We then identify additional high-level observables which are able to narrow this gap, and utilize LASSO regularization for feature selection to identify and rank the most relevant observables and provide further insights into the learning strategies used by the constituent-based neural networks. The final model contains only 31 high-level observables and is able to match the performance of the PFN and approximate the performance of the Transformer model to within 2%.
Article PDF
Similar content being viewed by others
Explore related subjects
Find the latest articles, discoveries, and news in related topics.Avoid common mistakes on your manuscript.
References
J.M. Butterworth, A.R. Davison, M. Rubin and G.P. Salam, Jet substructure as a new Higgs search channel at the LHC, Phys. Rev. Lett. 100 (2008) 242001 [arXiv:0802.2470] [INSPIRE].
A. Hook, E. Izaguirre, M. Lisanti and J.G. Wacker, High Multiplicity Searches at the LHC Using Jet Masses, Phys. Rev. D 85 (2012) 055029 [arXiv:1202.0558] [INSPIRE].
ATLAS collaboration, Performance of top-quark and W -boson tagging with ATLAS in Run 2 of the LHC, Eur. Phys. J. C 79 (2019) 375 [arXiv:1808.07858] [INSPIRE].
CMS collaboration, Identification techniques for highly boosted W bosons that decay into hadrons, JHEP 12 (2014) 017 [arXiv:1410.4227] [INSPIRE].
ATLAS collaboration, Measurement of jet-substructure observables in top quark, W boson and light jet production in proton-proton collisions at \( \sqrt{s} \) = 13 TeV with the ATLAS detector, JHEP 08 (2019) 033 [arXiv:1903.02942] [INSPIRE].
ATLAS collaboration, Measurement of the Lund Jet Plane Using Charged Particles in 13 TeV Proton-Proton Collisions with the ATLAS Detector, Phys. Rev. Lett. 124 (2020) 222002 [arXiv:2004.03540] [INSPIRE].
CMS collaboration, Measurement of jet substructure observables in \( t\overline{t} \) events from proton-proton collisions at \( \sqrt{s} \) = 13TeV, Phys. Rev. D 98 (2018) 092014 [arXiv:1808.07340] [INSPIRE].
CMS collaboration, Identification of heavy, energetic, hadronically decaying particles using machine-learning techniques, 2020 JINST 15 P06005 [arXiv:2004.08262] [INSPIRE].
CMS collaboration, Search for top squark production in fully-hadronic final states in proton-proton collisions at \( \sqrt{s} \) = 13 TeV, Phys. Rev. D 104 (2021) 052001 [arXiv:2103.01290] [INSPIRE].
ATLAS collaboration, Search for heavy diboson resonances in semileptonic final states in pp collisions at \( \sqrt{s} \) = 13 TeV with the ATLAS detector, Eur. Phys. J. C 80 (2020) 1165 [arXiv:2004.14636] [INSPIRE].
P.T. Komiske, E.M. Metodiev and J. Thaler, Energy flow polynomials: A complete linear basis for jet substructure, JHEP 04 (2018) 013 [arXiv:1712.07124] [INSPIRE].
A.J. Larkoski, G.P. Salam and J. Thaler, Energy Correlation Functions for Jet Substructure, JHEP 06 (2013) 108 [arXiv:1305.0007] [INSPIRE].
P.T. Komiske, E.M. Metodiev and M.D. Schwartz, Deep learning in color: towards automated quark/gluon jet discrimination, JHEP 01 (2017) 110 [arXiv:1612.01551] [INSPIRE].
P. Baldi, P. Sadowski and D. Whiteson, Searching for Exotic Particles in High-Energy Physics with Deep Learning, Nature Commun. 5 (2014) 4308 [arXiv:1402.4735] [INSPIRE].
L.G. Almeida, M. Backović, M. Cliche, S.J. Lee and M. Perelstein, Playing Tag with ANN: Boosted Top Identification with Pattern Recognition, JHEP 07 (2015) 086 [arXiv:1501.05968] [INSPIRE].
L. de Oliveira, M. Kagan, L. Mackey, B. Nachman and A. Schwartzman, Jet-images — deep learning edition, JHEP 07 (2016) 069 [arXiv:1511.05190] [INSPIRE].
P. Baldi, K. Bauer, C. Eng, P. Sadowski and D. Whiteson, Jet Substructure Classification in High-Energy Physics with Deep Neural Networks, Phys. Rev. D 93 (2016) 094034 [arXiv:1603.09349] [INSPIRE].
A.J. Larkoski, I. Moult and B. Nachman, Jet Substructure at the Large Hadron Collider: A Review of Recent Advances in Theory and Machine Learning, Phys. Rept. 841 (2020) 1 [arXiv:1709.04464] [INSPIRE].
G. Kasieczka, T. Plehn, M. Russell and T. Schell, Deep-learning Top Taggers or The End of QCD?, JHEP 05 (2017) 006 [arXiv:1701.08784] [INSPIRE].
P. Baldi, Deep Learning in Science, Cambridge University Press, Cambridge, U.K. (2021) [DOI].
T. Faucett, J. Thaler and D. Whiteson, Mapping Machine-Learned Physics into a Human-Readable Space, Phys. Rev. D 103 (2021) 036020 [arXiv:2010.11998] [INSPIRE].
L. Moore, K. Nordström, S. Varma and M. Fairbairn, Reports of My Demise Are Greatly Exaggerated: N-subjettiness Taggers Take On Jet Images, SciPost Phys. 7 (2019) 036 [arXiv:1807.04769] [INSPIRE].
A. Butter et al., The Machine Learning landscape of top taggers, SciPost Phys. 7 (2019) 014 [arXiv:1902.09914] [INSPIRE].
H. Qu and L. Gouskos, ParticleNet: Jet Tagging via Particle Clouds, Phys. Rev. D 101 (2020) 056019 [arXiv:1902.08570] [INSPIRE].
V. Mikuni and F. Canelli, Point cloud transformers applied to collider physics, Mach. Learn. Sci. Tech. 2 (2021) 035027 [arXiv:2102.05073] [INSPIRE].
CMS collaboration, Search for resonances decaying to three W bosons in the hadronic final state in proton-proton collisions at \( \sqrt{s} \) = 13 TeV, Phys. Rev. D 106 (2022) 012002 [arXiv:2112.13090] [INSPIRE].
J.A. Aguilar-Saavedra, F.R. Joaquim and J.F. Seabra, Mass Unspecific Supervised Tagging (MUST) for boosted jets, JHEP 03 (2021) 012 [Erratum ibid. 04 (2021) 133] [arXiv:2008.12792] [INSPIRE].
J.A. Aguilar-Saavedra, J.H. Collins and R.K. Mishra, A generic anti-QCD jet tagger, JHEP 11 (2017) 163 [arXiv:1709.01087] [INSPIRE].
C. Chen, Reconstruction and identification of H → WW* with high transverse momentum in the full hadronic final state, Phys. Rev. D 103 (2021) 033005 [arXiv:2012.02884] [INSPIRE].
J. Alwall et al., The automated computation of tree-level and next-to-leading order differential cross sections, and their matching to parton shower simulations, JHEP 07 (2014) 079 [arXiv:1405.0301] [INSPIRE].
C. Bierlich et al., A comprehensive guide to the physics and usage of PYTHIA 8.3, arXiv:2203.11601 [INSPIRE].
DELPHES 3 collaboration, DELPHES 3, A modular framework for fast simulation of a generic collider experiment, JHEP 02 (2014) 057 [arXiv:1307.6346] [INSPIRE].
M. Cacciari, G.P. Salam and G. Soyez, The anti-kt jet clustering algorithm, JHEP 04 (2008) 063 [arXiv:0802.1189] [INSPIRE].
M. Cacciari, G.P. Salam and G. Soyez, FastJet User Manual, Eur. Phys. J. C 72 (2012) 1896 [arXiv:1111.6097] [INSPIRE].
J. Thaler and K. Van Tilburg, Identifying Boosted Objects with N-subjettiness, JHEP 03 (2011) 015 [arXiv:1011.2268] [INSPIRE].
K. Datta and A. Larkoski, How Much Information is in a Jet?, JHEP 06 (2017) 073 [arXiv:1704.08249] [INSPIRE].
V. Nair and G.E. Hinton, Rectified linear units improve restricted boltzmann machines, in Proceedings of the 27th International Conference on International Conference on Machine Learning, ICML’10, Madison, U.S.A. June 2010, pp. 807–814.
P.T. Komiske, E.M. Metodiev and J. Thaler, Energy Flow Networks: Deep Sets for Particle Jets, JHEP 01 (2019) 121 [arXiv:1810.05165] [INSPIRE].
A. Vaswani et al., Attention is all you need, arXiv:1706.03762.
J. Collado, K. Bauer, E. Witkowski, T. Faucett, D. Whiteson and P. Baldi, Learning to isolate muons, JHEP 21 (2020) 200 [arXiv:2102.02278] [INSPIRE].
J. Devlin, M.-W. Chang, K. Lee and K. Toutanova, Bert: Pre-training of deep bidirectional transformers for language understanding, arXiv:1810.04805.
A. Radford et al., Language models are unsupervised multitask learners, OpenAI blog 1 (2019) 9.
M. Lewis et al., Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension, arXiv:1910.13461.
M.J. Fenton, A. Shmakov, T.-W. Ho, S.-C. Hsu, D. Whiteson and P. Baldi, Permutationless many-jet event reconstruction with symmetry preserving attention networks, Phys. Rev. D 105 (2022) 112008 [arXiv:2010.09206] [INSPIRE].
A. Shmakov, M.J. Fenton, T.-W. Ho, S.-C. Hsu, D. Whiteson and P. Baldi, SPANet: Generalized permutationless set assignment for particle physics using symmetry preserving attention, SciPost Phys. 12 (2022) 178 [arXiv:2106.03898] [INSPIRE].
J. Collado, J.N. Howard, T. Faucett, T. Tong, P. Baldi and D. Whiteson, Learning to identify electrons, Phys. Rev. D 103 (2021) 116028 [arXiv:2011.01984] [INSPIRE].
J. Gallicchio and M.D. Schwartz, Quark and Gluon Jet Substructure, JHEP 04 (2013) 090 [arXiv:1211.7038] [INSPIRE].
R. Tibshirani, Regression shrinkage and selection via the lasso, J. Roy. Statist. Soc. 58 (1996) 267.
A. Hoffmann, R. Kwok and P. Compton, Using subclasses to improve classification learning, in Machine Learning: ECML 2001, L. De Raedt and P. Flach, eds., Springer Berlin, Berlin, Heidelberg, Germany (2001), pp. 203–213 [DOI].
Y. Luo, Can subclasses help a multiclass learning problem?, in 2008 IEEE Intelligent Vehicles Symposium, 4–6 June 2008, pp. 214–219 [DOI].
D.P. Kingma and J. Ba, Adam: A Method for Stochastic Optimization, (2014) [arXiv:1412.6980] [INSPIRE].
L. Hertel, J. Collado, P. Sadowski, J. Ott and P. Baldi, Sherpa: Robust hyperparameter optimization for machine learning, SoftwareX (2020), arXiv:2005.04048v1.
A. Paszke et al., Pytorch: An imperative style, high-performance deep learning library, in Advances in Neural Information Processing Systems 32, H. Wallach, H. Larochelle, A. Beygelzimer, F. d’Alché-Buc, E. Fox and R. Garnett eds., Curran Associates Inc., Red Hook, U.S.A. (2019), pp. 8024–8035 [http://papers.neurips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library.pdf].
F. Chollet et al., Keras, https://keras.io.
M. Abadi et al., TensorFlow: Large-scale machine learning on heterogeneous systems, arXiv:1603.04467.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
ArXiv ePrint: 2202.00723
Rights and permissions
Open Access . This article is distributed under the terms of the Creative Commons Attribution License (CC-BY 4.0), which permits any use, distribution and reproduction in any medium, provided the original author(s) and source are credited.
About this article
Cite this article
Lu, Y., Romero, A., Fenton, M.J. et al. Resolving extreme jet substructure. J. High Energ. Phys. 2022, 46 (2022). https://doi.org/10.1007/JHEP08(2022)046
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/JHEP08(2022)046