Abstract
An important class of techniques for resonant anomaly detection in high energy physics builds models that can distinguish between reference and target datasets, where only the latter has appreciable signal. Such techniques, including Classification Without Labels (CWoLa) and Simulation Assisted Likelihood-free Anomaly Detection (SALAD) rely on a single reference dataset. They cannot take advantage of commonly-available multiple datasets and thus cannot fully exploit available information. In this work, we propose generalizations of CWoLa and SALAD for settings where multiple reference datasets are available, building on weak supervision techniques. We demonstrate improved performance in a number of settings with realistic and synthetic data. As an added benefit, our generalizations enable us to provide finite-sample guarantees, improving on existing asymptotic analyses.
Article PDF
Similar content being viewed by others
Avoid common mistakes on your manuscript.
References
H.E.P.M.L. community, A living review of machine learning for particle physics, https://iml-wg.github.io/HEPML-LivingReview/.
G. Karagiorgi et al., Machine learning in the search for new fundamental physics, arXiv:2112.03769 [INSPIRE].
G. Kasieczka et al., The LHC olympics 2020 a community challenge for anomaly detection in high energy physics, Rept. Prog. Phys. 84 (2021) 124201 [arXiv:2101.08320] [INSPIRE].
T. Aarrestad et al., The dark machines anomaly score challenge: benchmark data and model independent event classification for the Large Hadron Collider, SciPost Phys. 12 (2022) 043 [arXiv:2105.14027] [INSPIRE].
ATLAS collaboration, Dijet resonance search with weak supervision using \( \sqrt{s} \) = 13 TeV pp collisions in the ATLAS detector, Phys. Rev. Lett. 125 (2020) 131801 [arXiv:2005.02983] [INSPIRE].
ATLAS Collaboration collaboration, Anomaly detection search for new resonances decaying into a Higgs boson and a generic new particle X in hadronic final states using \( \sqrt{s} \) = 13 TeV pp collisions with the ATLAS detector, ATLAS-CONF-2022-045, CERN, Geneva, Switzerland (2022).
G. Kasieczka et al., Anomaly detection under coordinate transformations, Phys. Rev. D 107 (2023) 015009 [arXiv:2209.06225] [INSPIRE].
G. Kasieczka, B. Nachman and D. Shih, New methods and datasets for group anomaly detection from fundamental physics, in the proceedings of the Conference on knowledge discovery and data mining, (2021) [arXiv:2107.02821] [INSPIRE].
J.H. Collins, K. Howe and B. Nachman, Anomaly detection for resonant new physics with machine learning, Phys. Rev. Lett. 121 (2018) 241803 [arXiv:1805.02664] [INSPIRE].
J.H. Collins, K. Howe and B. Nachman, Extending the search for new resonances with machine learning, Phys. Rev. D 99 (2019) 014038 [arXiv:1902.02634] [INSPIRE].
R.T. D’Agnolo and A. Wulzer, Learning new physics from a machine, Phys. Rev. D 99 (2019) 015014 [arXiv:1806.02350] [INSPIRE].
R.T. D’Agnolo et al., Learning multivariate new physics, Eur. Phys. J. C 81 (2021) 89 [arXiv:1912.12155] [INSPIRE].
K. Benkendorfer, L.L. Pottier and B. Nachman, Simulation-assisted decorrelation for resonant anomaly detection, Phys. Rev. D 104 (2021) 035003 [arXiv:2009.02205] [INSPIRE].
A. Andreassen, B. Nachman and D. Shih, Simulation assisted likelihood-free anomaly detection, Phys. Rev. D 101 (2020) 095004 [arXiv:2001.05001] [INSPIRE].
B. Nachman and D. Shih, Anomaly detection with density estimation, Phys. Rev. D 101 (2020) 075042 [arXiv:2001.04990] [INSPIRE].
O. Amram and C.M. Suarez, Tag N’ Train: a technique to train improved classifiers on unlabeled data, JHEP 01 (2021) 153 [arXiv:2002.12376] [INSPIRE].
A. Hallin et al., Classifying anomalies through outer density estimation, Phys. Rev. D 106 (2022) 055006 [arXiv:2109.00546] [INSPIRE].
J.A. Raine, S. Klein, D. Sengupta and T. Golling, CURTAINs for your sliding window: constructing unobserved regions by transforming adjacent intervals, Front. Big Data 6 (2023) 899345 [arXiv:2203.09470] [INSPIRE].
R.T. d’Agnolo et al., Learning new physics from an imperfect machine, Eur. Phys. J. C 82 (2022) 275 [arXiv:2111.13633] [INSPIRE].
P. Chakravarti, M. Kuusela, J. Lei and L. Wasserman, Model-independent detection of new physics signals using interpretable semi-supervised classifier tests, arXiv:2102.07679 [INSPIRE].
B.M. Dillon, R. Mastandrea and B. Nachman, Self-supervised anomaly detection for new physics, Phys. Rev. D 106 (2022) 056005 [arXiv:2205.10380] [INSPIRE].
M. Letizia et al., Learning new physics efficiently with nonparametric methods, Eur. Phys. J. C 82 (2022) 879 [arXiv:2204.02317] [INSPIRE].
K. Krzyżańska and B. Nachman, Simulation-based anomaly detection for multileptons at the LHC, JHEP 01 (2023) 061 [arXiv:2203.09601] [INSPIRE].
S. Alvi, C.W. Bauer and B. Nachman, Quantum anomaly detection for collider physics, JHEP 02 (2023) 220 [arXiv:2206.08391] [INSPIRE].
T. Sjostrand, S. Mrenna and P.Z. Skands, A brief introduction to PYTHIA 8.1, Comput. Phys. Commun. 178 (2008) 852 [arXiv:0710.3820] [INSPIRE].
J. Bellm et al., Herwig 7.0/Herwig++ 3.0 release note, Eur. Phys. J. C 76 (2016) 196 [arXiv:1512.01178] [INSPIRE].
Sherpa collaboration, Event generation with Sherpa 2.2, SciPost Phys. 7 (2019) 034 [arXiv:1905.09127] [INSPIRE].
E.M. Metodiev, B. Nachman and J. Thaler, Classification without labels: learning from mixed samples in high energy physics, JHEP 10 (2017) 174 [arXiv:1708.02949] [INSPIRE].
J. Neyman and E.S. Pearson, On the problem of the most efficient tests of statistical hypotheses, Phil. Trans. Roy. Soc. Lond. A 231 (1933) 289 [INSPIRE].
A. Ratner et al., Snorkel: rapid training data creation with weak supervision, in the Proceedings of the of the 44th international conference on Very Large Data Bases (VLDB), Rio de Janeiro, Brazil (2018).
D.Y. Fu et al., Fast and three-rious: speeding up weak supervision with triplet methods, in International conference on machine learning, (2020) [arXiv:2002.11955].
A. Ratner et al., Data programming: creating large training sets, quickly, in the Proceedings of the of the 30th International Conference on Neural Information Processing Systems, Red Hook, NY, U.S.A. (2016), p. 3574.
A. Ratner et al., Training complex models with multi-task weak supervision, in the Proceedings of the of the AAAI Conference on Artificial Intelligence, (2019).
P. Varma et al., Learning dependency structures for weak supervision models, in the Proceedings of the of the 36th International Conference on Machine Learning, (2019).
M.J. Wainwright and M.I. Jordan, Graphical models, exponential families, and variational inference, Foundations and Trends® in Machine Learning 1 (2007) 1.
J. Thaler and K. Van Tilburg, Identifying boosted objects with N -subjettiness, JHEP 03 (2011) 015 [arXiv:1011.2268] [INSPIRE].
J. Thaler and K. Van Tilburg, Maximizing boosted top identification by minimizing N -subjettiness, JHEP 02 (2012) 093 [arXiv:1108.2701] [INSPIRE].
B. Nachman and J. Thaler, Learning from many collider events at once, Phys. Rev. D 103 (2021) 116013 [arXiv:2101.07263] [INSPIRE].
T. Hastie, R. Tibshirani and J. Friedman, The elements of statistical learning, Springer, New York, NY, U.S.A. (2009) [https://doi.org/10.1007/978-0-387-84858-7].
M. Sugiyama, T. Suzuki and T. Kanamori, Density ratio estimation in machine learning, Cambridge University Press, Cambridge, U.K. (2012) [https://doi.org/10.1017/cbo9781139035613].
K. Cranmer, J. Pavez and G. Louppe, Approximating likelihood ratios with calibrated discriminative classifiers, arXiv:1506.02169 [INSPIRE].
S. Dasgupta and P.M. Long, Boosting with diverse base classifiers, in the Proceedings of the Learning Theory and Kernel Machines, Berlin, Heidelberg, Germany (2003), p. 273.
C. Cortes, Y. Mansour and M. Mohri, Learning bounds for importance weighting, in the Proceedings of the Advances in Neural Information Processing Systems, Curran Associates Inc. (2010).
T. Ma, Lecture notes for machine learning theory (CS229M/STATS214), https://web.stanford.edu/class/stats214/ June 2022
P.L. Bartlett and S. Mendelson, Rademacher and Gaussian complexities: risk bounds and structural results, J. Mach. Learn. Res. 3 (2002) 463.
M. Mohri, A. Rostamizadeh and A. Talwalkar, Foundations of machine learning, MIT Press, Cambridge, MA, U.S.A. (2018).
A. Paszke et al., PyTorch: an imperative style, high-performance deep learning library, in Advances in neural information processing systems 32, Curran Associates Inc. (2019), p. 8024.
F. Chollet et al., Keras, https://github.com/fchollet/keras.
Acknowledgments
We thank David Shih and Jesse Thaler for useful discussions and comments about the manuscript. BN was supported by the Department of Energy, Office of Science under contract number DE-AC02-05CH11231. FS is grateful for the support of the NSF under CCF2106707 and the Wisconsin Alumni Research Foundation (WARF). We gratefully acknowledge the support of NIH under No. U54EB020405 (Mobilize), NSF under Nos. CCF1763315 (Beyond Sparsity), CCF1563078 (Volume to Velocity), and 1937301 (RTML); ARL under No. W911NF-21-2-0251 (Interactive Human-AI Teaming); ONR under No. N000141712266 (UnifyingWeak Supervision); ONR N00014-20-1-2480: Understanding and Applying Non-Euclidean Geometry in Machine Learning; N000142012275 (NEPTUNE); NXP, Xilinx, LETI-CEA, Intel, IBM, Microsoft, NEC, Toshiba, TSMC, ARM, Hitachi, BASF, Accenture, Ericsson, Qualcomm, Analog Devices, Google Cloud, Salesforce, Total, the HAI-GCP Cloud Credits for Research program, the Stanford Data Science Initiative (SDSI), and members of the Stanford DAWN project: Facebook, Google, and VMWare. The U.S. Government is authorized to reproduce and distribute reprints for Governmental purposes notwithstanding any copyright notation thereon. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views, policies, or endorsements, either expressed or implied, of NIH, ONR, or the U.S. Government.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
ArXiv ePrint: 2212.10579
Rights and permissions
Open Access . This article is distributed under the terms of the Creative Commons Attribution License (CC-BY 4.0), which permits any use, distribution and reproduction in any medium, provided the original author(s) and source are credited.
About this article
Cite this article
Chen, M.F., Nachman, B. & Sala, F. Resonant anomaly detection with multiple reference datasets. J. High Energ. Phys. 2023, 188 (2023). https://doi.org/10.1007/JHEP07(2023)188
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/JHEP07(2023)188