Abstract
Deep learning methods have been increasingly adopted to study jets in particle physics. Since symmetry-preserving behavior has been shown to be an important factor for improving the performance of deep learning in many applications, Lorentz group equivariance — a fundamental spacetime symmetry for elementary particles — has recently been incorporated into a deep learning model for jet tagging. However, the design is computationally costly due to the analytic construction of high-order tensors. In this article, we introduce LorentzNet, a new symmetry-preserving deep learning model for jet tagging. The message passing of LorentzNet relies on an efficient Minkowski dot product attention. Experiments on two representative jet tagging benchmarks show that LorentzNet achieves the best tagging performance and improves significantly over existing state-of-the-art algorithms. The preservation of Lorentz symmetry also greatly improves the efficiency and generalization power of the model, allowing LorentzNet to reach highly competitive performance when trained on only a few thousand jets.
Article PDF
Similar content being viewed by others
Explore related subjects
Find the latest articles, discoveries, and news in related topics.Avoid common mistakes on your manuscript.
References
R. Kogler et al., Jet Substructure at the Large Hadron Collider: Experimental Review, Rev. Mod. Phys. 91 (2019) 045003 [arXiv:1803.06991] [INSPIRE].
A. Abdesselam et al., Boosted Objects: A Probe of Beyond the Standard Model Physics, Eur. Phys. J. C 71 (2011) 1661 [arXiv:1012.5412] [INSPIRE].
A. Altheimer et al., Jet Substructure at the Tevatron and LHC: New results, new tools, new benchmarks, J. Phys. G 39 (2012) 063001 [arXiv:1201.0008] [INSPIRE].
A. Altheimer et al., Boosted Objects and Jet Substructure at the LHC. Report of BOOST2012, held at IFIC Valencia, 23rd-27th of July 2012, Eur. Phys. J. C 74 (2014) 2792 [arXiv:1311.2708] [INSPIRE].
D. Adams et al., Towards an Understanding of the Correlations in Jet Substructure, Eur. Phys. J. C 75 (2015) 409 [arXiv:1504.00679] [INSPIRE].
S. Marzani, G. Soyez and M. Spannowsky, Looking inside jets: an introduction to jet substructure and boosted-object phenomenology, vol. 958, Springer, Cham, Germany (2019) [DOI] [arXiv:1901.10342] [INSPIRE].
A. J. Larkoski, I. Moult and B. Nachman, Jet Substructure at the Large Hadron Collider: A Review of Recent Advances in Theory and Machine Learning, Phys. Rept. 841 (2020) 1 [arXiv:1709.04464] [INSPIRE].
P. T. Komiske, E. M. Metodiev and J. Thaler, Energy flow polynomials: A complete linear basis for jet substructure, JHEP 04 (2018) 013 [arXiv:1712.07124] [INSPIRE].
K. Datta and A. Larkoski, How Much Information is in a Jet?, JHEP 06 (2017) 073 [arXiv:1704.08249] [INSPIRE].
K. Datta and A. J. Larkoski, Novel Jet Observables from Machine Learning, JHEP 03 (2018) 086 [arXiv:1710.01305] [INSPIRE].
K. Datta, A. Larkoski and B. Nachman, Automating the Construction of Jet Observables with Machine Learning, Phys. Rev. D 100 (2019) 095016 [arXiv:1902.07180] [INSPIRE].
A. Chakraborty, S. H. Lim and M. M. Nojiri, Interpretable deep learning for two-prong jet classification with jet spectra, JHEP 07 (2019) 135 [arXiv:1904.02092] [INSPIRE].
A. Chakraborty, S. H. Lim, M. M. Nojiri and M. Takeuchi, Neural Network-based Top Tagger with Two-Point Energy Correlations and Geometry of Soft Emissions, JHEP 07 (2020) 111 [arXiv:2003.11787] [INSPIRE].
J. Cogan, M. Kagan, E. Strauss and A. Schwarztman, Jet-Images: Computer Vision Inspired Techniques for Jet Tagging, JHEP 02 (2015) 118 [arXiv:1407.5675] [INSPIRE].
L. G. Almeida, M. Backović, M. Cliche, S. J. Lee and M. Perelstein, Playing Tag with ANN: Boosted Top Identification with Pattern Recognition, JHEP 07 (2015) 086 [arXiv:1501.05968] [INSPIRE].
P. Baldi, K. Bauer, C. Eng, P. Sadowski and D. Whiteson, Jet Substructure Classification in High-Energy Physics with Deep Neural Networks, Phys. Rev. D 93 (2016) 094034 [arXiv:1603.09349] [INSPIRE].
L. de Oliveira, M. Kagan, L. Mackey, B. Nachman and A. Schwartzman, Jet-images — deep learning edition, JHEP 07 (2016) 069 [arXiv:1511.05190] [INSPIRE].
J. Barnard, E. N. Dawe, M. J. Dolan and N. Rajcic, Parton Shower Uncertainties in Jet Substructure Analyses with Deep Neural Networks, Phys. Rev. D 95 (2017) 014018 [arXiv:1609.00607] [INSPIRE].
P. T. Komiske, E. M. Metodiev and M. D. Schwartz, Deep learning in color: towards automated quark/gluon jet discrimination, JHEP 01 (2017) 110 [arXiv:1612.01551] [INSPIRE].
G. Kasieczka, T. Plehn, M. Russell and T. Schell, Deep-learning Top Taggers or The End of QCD?, JHEP 05 (2017) 006 [arXiv:1701.08784] [INSPIRE].
S. Macaluso and D. Shih, Pulling Out All the Tops with Computer Vision and Deep Learning, JHEP 10 (2018) 121 [arXiv:1803.00107] [INSPIRE].
S. Choi, S. J. Lee and M. Perelstein, Infrared Safety of a Neural-Net Top Tagging Algorithm, JHEP 02 (2019) 132 [arXiv:1806.01263] [INSPIRE].
F. A. Dreyer, G. P. Salam and G. Soyez, The Lund Jet Plane, JHEP 12 (2018) 064 [arXiv:1807.04758] [INSPIRE].
J. Lin, M. Freytsis, I. Moult and B. Nachman, Boosting \( H\to b\overline{b} \) with Machine Learning, JHEP 10 (2018) 101 [arXiv:1807.10768] [INSPIRE].
Y.-L. Du et al., Identifying the nature of the QCD transition in relativistic collision of heavy nuclei with deep learning, Eur. Phys. J. C 80 (2020) 516 [arXiv:1910.11530] [INSPIRE].
J. Li, T. Li and F.-Z. Xu, Reconstructing boosted Higgs jets from event image segmentation, JHEP 04 (2021) 156 [arXiv:2008.13529] [INSPIRE].
J. Filipek, S.-C. Hsu, J. Kruper, K. Mohan and B. Nachman, Identifying the Quantum Properties of Hadronic Resonances using Machine Learning, arXiv:2105.04582 [INSPIRE].
K. Fraser and M. D. Schwartz, Jet Charge and Machine Learning, JHEP 10 (2018) 093 [arXiv:1803.08066] [INSPIRE].
CMS collaboration, Identification of heavy, energetic, hadronically decaying particles using machine-learning techniques, 2020 JINST 15 P06005 [arXiv:2004.08262] [INSPIRE].
D. Guest, J. Collado, P. Baldi, S.-C. Hsu, G. Urban and D. Whiteson, Jet Flavor Classification in High-Energy Physics with Deep Neural Networks, Phys. Rev. D 94 (2016) 112002 [arXiv:1607.08633] [INSPIRE].
J. Pearkes, W. Fedorko, A. Lister and C. Gay, Jet Constituents for Deep Neural Network Based Top Quark Tagging, arXiv:1704.02124 [INSPIRE].
A. Butter, G. Kasieczka, T. Plehn and M. Russell, Deep-learned Top Tagging with a Lorentz Layer, SciPost Phys. 5 (2018) 028 [arXiv:1707.08966] [INSPIRE].
G. Kasieczka, N. Kiefer, T. Plehn and J. M. Thompson, Quark-Gluon Tagging: Machine Learning vs Detector, SciPost Phys. 6 (2019) 069 [arXiv:1812.09223] [INSPIRE].
S. Egan, W. Fedorko, A. Lister, J. Pearkes and C. Gay, Long Short-Term Memory (LSTM) networks with jet constituents for boosted top tagging at the LHC, arXiv:1711.09059 [INSPIRE].
E. Bols, J. Kieseler, M. Verzetti, M. Stoye and A. Stakia, Jet Flavour Classification Using DeepJet, 2020 JINST 15 P12012 [arXiv:2008.10519] [INSPIRE].
P. T. Komiske, E. M. Metodiev and J. Thaler, Energy Flow Networks: Deep Sets for Particle Jets, JHEP 01 (2019) 121 [arXiv:1810.05165] [INSPIRE].
H. Qu and L. Gouskos, ParticleNet: Jet Tagging via Particle Clouds, Phys. Rev. D 101 (2020) 056019 [arXiv:1902.08570] [INSPIRE].
E. A. Moreno et al., JEDI-net: a jet identification algorithm based on interaction networks, Eur. Phys. J. C 80 (2020) 58 [arXiv:1908.05318] [INSPIRE].
E. A. Moreno et al., Interaction networks for the identification of boosted \( H\to b\overline{b} \) decays, Phys. Rev. D 102 (2020) 012010 [arXiv:1909.12285] [INSPIRE].
V. Mikuni and F. Canelli, ABCNet: An attention-based method for particle tagging, Eur. Phys. J. Plus 135 (2020) 463 [arXiv:2001.05311] [INSPIRE].
E. Bernreuther, T. Finke, F. Kahlhoefer, M. Krämer and A. Mück, Casting a graph net to catch dark showers, SciPost Phys. 10 (2021) 046 [arXiv:2006.08639] [INSPIRE].
J. Guo, J. Li, T. Li and R. Zhang, Boosted Higgs boson jet reconstruction via a graph neural network, Phys. Rev. D 103 (2021) 116025 [arXiv:2010.05464] [INSPIRE].
M. J. Dolan and A. Ore, Equivariant Energy Flow Networks for Jet Tagging, Phys. Rev. D 103 (2021) 074022 [arXiv:2012.00964] [INSPIRE].
V. Mikuni and F. Canelli, Point cloud transformers applied to collider physics, Mach. Learn. Sci. Tech. 2 (2021) 035027 [arXiv:2102.05073] [INSPIRE].
P. Konar, V. S. Ngairangbam and M. Spannowsky, Energy-weighted message passing: an infra-red and collinear safe graph neural network algorithm, JHEP 02 (2022) 060 [arXiv:2109.14636] [INSPIRE].
C. Shimmin, Particle Convolution for High Energy Physics, arXiv:2107.02908 [INSPIRE].
G. Louppe, K. Cho, C. Becot and K. Cranmer, QCD-Aware Recursive Neural Networks for Jet Physics, JHEP 01 (2019) 057 [arXiv:1702.00748] [INSPIRE].
T. Cheng, Recursive Neural Networks in Quark/Gluon Tagging, Comput. Softw. Big Sci. 2 (2018) 3 [arXiv:1711.02633] [INSPIRE].
F. A. Dreyer and H. Qu, Jet tagging in the Lund plane with graph networks, JHEP 03 (2021) 052 [arXiv:2012.08526] [INSPIRE].
F. Dreyer, G. Soyez and A. Takacs, Quarks and gluons in the Lund plane, arXiv:2112.09140 [INSPIRE].
T. Cohen and M. Welling, Group equivariant convolutional networks, in International conference on machine learning, pp. 2990–2999, PMLR (2016) [arXiv:1602.07576].
M. Zaheer, S. Kottur, S. Ravanbakhsh, B. Poczos, R. R. Salakhutdinov and A. J. Smola, Deep sets, Adv. Neural Inf. Process. Syst. 30 (2017) [arXiv:1703.06114].
M. Finzi, S. Stanton, P. Izmailov and A. G. Wilson, Generalizing convolutional neural networks for equivariance to lie groups on arbitrary continuous data, in International Conference on Machine Learning, pp. 3165–3176, PMLR (2020) [arXiv:2002.12880].
G. Kanwar et al., Equivariant flow-based sampling for lattice gauge theory, Phys. Rev. Lett. 125 (2020) 121601 [arXiv:2003.06413] [INSPIRE].
D. Boyda et al., Sampling using SU(N) gauge equivariant flows, Phys. Rev. D 103 (2021) 074504 [arXiv:2008.05456] [INSPIRE].
V. G. Satorras, E. Hoogeboom and M. Welling, E(n) equivariant graph neural networks, arXiv:2102.09844.
M. J. Hutchinson, C. Le Lan, S. Zaidi, E. Dupont, Y. W. Teh and H. Kim, Lietransformer: Equivariant self-attention for lie groups, in International Conference on Machine Learning, pp. 4533–4543, PMLR (2021) [arXiv:2012.10885].
S. Batzner et al., E(3)-equivariant graph neural networks for data-efficient and accurate interatomic potentials, arXiv:2101.03164.
K. Lenc and A. Vedaldi, Understanding image representations by measuring their equivariance and equivalence, in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 991–999 (2015) [DOI].
M. Erdmann, E. Geiser, Y. Rath and M. Rieger, Lorentz Boost Networks: Autonomous Physics-Inspired Feature Engineering, 2019 JINST 14 P06006 [arXiv:1812.09722] [INSPIRE].
A. Bogatskiy, B. Anderson, J. Offermann, M. Roussi, D. Miller and R. Kondor, Lorentz group equivariant neural network for particle physics, in International Conference on Machine Learning, pp. 992–1002, PMLR (2020) [arXiv:2006.04780].
F. Scarselli, M. Gori, A. C. Tsoi, M. Hagenbuchner and G. Monfardini, The graph neural network model, IEEE Trans. Neural Networks 20 (2008) 61.
J. Gilmer, S. S. Schoenholz, P. F. Riley, O. Vinyals and G. E. Dahl, Neural message passing for quantum chemistry, in International conference on machine learning, pp. 1263–1272, PMLR (2017) [arXiv:1704.01212].
S. Villar, D. Hogg, K. Storey-Fisher, W. Yao and B. Blum-Smith, Scalars are universal: Equivariant machine learning, structured like classical physics, Adv. Neural Inf. Process. Syst. 34 (2021) [arXiv:2106.06610].
I. Loshchilov and F. Hutter, Decoupled weight decay regularization, in International Conference on Learning Representations, (2018) [arXiv:1711.05101].
I. Loshchilov and F. Hutter, SGDR: Stochastic gradient descent with warm restarts, arXiv:1608.03983.
G. Kasieczka, T. Plehn and M. Russel, Top quark tagging reference dataset, https://doi.org/10.5281/zenodo.2603256 (2019).
S. Xie, R. Girshick, P. Dollár, Z. Tu and K. He, Aggregated residual transformations for deep neural networks, in Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1492–1500 (2017) [DOI].
CMS collaboration, Boosted jet identification using particle candidates and deep neural networks, CMS-DP-2017-049 (2017).
W. Du et al., Equivariant vector field network for many-body system modeling, arXiv:2110.14811.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
ArXiv ePrint: 2201.08187
This work was done when the first author was visiting Microsoft Research Asia. (Shiqi Gong)
Rights and permissions
Open Access . This article is distributed under the terms of the Creative Commons Attribution License (CC-BY 4.0), which permits any use, distribution and reproduction in any medium, provided the original author(s) and source are credited.
About this article
Cite this article
Gong, S., Meng, Q., Zhang, J. et al. An efficient Lorentz equivariant graph neural network for jet tagging. J. High Energ. Phys. 2022, 30 (2022). https://doi.org/10.1007/JHEP07(2022)030
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/JHEP07(2022)030