Liquid argon light collection and veto modeling in GERDA Phase II

The ability to detect liquid argon scintillation light from within a densely packed high-purity germanium detector array allowed the Gerda experiment to reach an exceptionally low background rate in the search for neutrinoless double beta decay of \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${}^{76}$$\end{document}76Ge. Proper modeling of the light propagation throughout the experimental setup, from any origin in the liquid argon volume to its eventual detection by the novel light read-out system, provides insight into the rejection capability and is a necessary ingredient to obtain robust background predictions. In this paper, we present a model of the Gerda liquid argon veto, as obtained by Monte Carlo simulations and constrained by calibration data, and highlight its application for background decomposition.


Introduction
Provided with an array of germanium detectors, made from isotopically enriched high-purity germanium (HPGe) material suspended in a clean liquid argon (LAr) bath, the Germanium Detector Array (Gerda) experiment set out to probe the neutrino's particle nature in a search for the neutrinoless double beta (0νββ) decay of 76 Ge [1].The ability to detect scintillation light emerging from coincident energy depositions in the LAr, combined with pulse shape discrimination (PSD) techniques [2], allowed to cut the background level of the second phase (Phase II) to a record low [3].No signal was found, which translates into one of the most stringent lower limits on the half-life of the 0νββ decay of 76 Ge at 1.8 • 10 26 yr at 90 % C.L. [4].
Based on dedicated Monte Carlo simulations of the scintillation light propagation, the model of the LAr veto rejection grants insight into the light collection from various regions of the highly heterogeneous setup.The full methodology and its first application are described in this document, which is structured as follows: Sec. 2 offers a brief description of the Gerda instrumentation, focusing on the Phase II light read-out system.In Sec. 3 the connection between photon detection probabilities and event rejection is made.Sec. 4 summarizes the Monte Carlo implementation and the chosen optia correspondence: gerda-eb@mpi-hd.mpg.deb also at: NRNU MEPhI, Moscow, Russia c present address: Duke University, Durham, NC USA d also at: Moscow Inst. of Physics and Technology, Russia e present address: Leibniz-Institut für Kristallzüchtung, Berlin, Germany f also at: Dubna State University, Dubna, Russia g present address: Nuclear Science Division, Berkeley, USA h present address: Université Paris-Saclay, CNRS/IN2P3, IJ-CLab, 91405 Orsay, France cal properties, while Sec. 5 describes the tuning of the model parameters on calibration data.In Sec. 6 photon detection probability maps are introduced, and in Sec.7 their application for background decomposition is highlighted.In Sec. 8 conclusions are drawn.

Instrumentation
The Gerda experimental site was the Hall A of the INFN Laboratori Nazionali del Gran Sasso (LNGS) underground laboratory in central Italy.Equipped with a large-scale shielding infrastructure -a 64 m 3 cryostat inside a 590 m 3 water tank -Gerda enclosed a low-background LAr environment, which from December 2015 to November 2019 gave home to the heart of Phase II: 40, later 41, HPGe detectors in a 7-string array configuration, surrounded by a light read-out instrumentation.The veto design comprised two subsystems: low-activity photomultiplier tubes (PMTs) [5] and wavelength-shifting (WLS) fibers coupled to silicon photomultipliers (SiPMs) [6].The latter was upgraded in spring 2018.
The 3" Hamamatsu R11065-20 Mod PMTs were selected for their performance at LAr temperature and enhanced radiopurity [7,8].Still, they contributed significantly to the background within the inner Phase II setup and were thus placed at >1 m from any HPGe detector, giving -together with the limited cryostat entrance width -the LAr instrumentation its elongated cylindrical shape.With space for support and calibration sources to enter, 3 off-center groups of 3 PMTs each were installed on top, whereas the bottom plate held 7 centrally mounted PMTs.Each PMT's entrance window was covered with tetraphenyl butadiene (TPB) embedded in polystyrene, in order to shift the incident vacuum-ultraviolet (VUV) scintillation light from LAr to a detectable wavelength.On the inside, the horizontal copper support plates were covered with a highly reflective TPB-painted VM2000 multi-layer polymer, whereas lateral guidance of light towards the top/bottom was ensured by a TPB dip-coated diffuse-reflecting Tetratex ® polytetrafluoroethylene (PTFE) foil [9], stitched to the 100 μm-thin copper shrouds.A sketch of the setup is shown in Fig. 1.
With typical SiPMs having a photo-sensitive area of O(1) cm 2 , large-scale installations of >1 m 2 photo coverage still represent a technological challenge [10].Nonetheless, coupled to WLS fibers of <0.1 mBq/kg activity in both 228 Th and 226 Ra, which serve as radiopure light collectors [11], the detection power of a single device is largely enhanced.The active chip size of the KETEK PM33100 SiPMs is 3 × 3 mm 2 , they feature 100 μm micro cell pitch and were purchased "in die", i. e. without packaging, allowing for a custom low-activity housing.Each of the 15 channels was comprised of 6 SiPMs, connected in parallel on copper-laminated PTFE holders and cast into optical cement, amounting to a total active surface of 8.1 cm 2 .The doubly-cladded BCF-91A fibers of square-shaped 1 × 1 mm 2 cross section, were covered with TPB by evaporation, routed vertically to cover the central veto section, bent by 180°at the bottom and coupled to different SiPMs on both top ends.Guidance of the individual fibers was ensured by micromachined copper holders, attempting to keep them at a 45°rotation, facing their full √ 2 mm-diagonal towards the center.The total length of the 405 fibers was about 730 m.
Each of the 40 cm-long HPGe strings was enclosed in a nylon "mini shroud", transparent to visible light, covered on both sides with TPB.It provided a mechanical barrier that limited the accumulation of 42 K ionsa progeny of cosmogenic 42 Ar -on the HPGe detector surfaces [12].
The data acquisition of the entire array, based on SIS3301 Struck [13] FADCs, including the LAr veto photo sensors, was triggered once the signal of a single HPGe detector exceeded a pre-set online threshold.No independent trigger on the light read-out was implemented.The veto condition was evaluated offline, allowing for time-dependent channel-specific thresholds right above the respective noise pedestals and an anticoincidence window that takes into account the characteristic scintillation emission timing as well as the HPGe detector signal formation dynamics.

Photon detection probabilities
Upon interaction of ionizing radiation, ultra-pure LAr scintillates with a light yield of 40 photons/keV [14].There is an ongoing discussion whether this number could be smaller [15], but in any case, the actual light output is strongly reduced in the presence of trace contaminants and a priori not precisely known for many experiments, including Gerda.An estimate based on the measured triplet lifetime of the argon excimer state of about 1.0 μs [3], limits the Gerda light yield to <71 % of the nominal pure-argon value, or <28 photons/keV. 1iven a light yield L of this order, the number of primary VUV photons produced in a typical Gerda background event can be enormous.Coincident energy depositions due to β and γ interactions in LAr (e. g. from 228 Th or 238 U trace impurities) frequently reach MeVenergies.The computational effort to track all O(10 4 ) optical photons represents a challenge, especially when considering the feedback between rejection power and required statistics -the larger the coincident energy release, the larger the suppression, the larger the statistics required to obtain a proper prediction of the HPGe spectrum after veto application.However, there is a workaround for this problem: the light propagation can be separated from the simulations that provide the energy depositions in the LAr.
The number of primary photons n generated from a single energy deposition (E, x) in the LAr, follows a Poisson distribution P n (λ) = λ n e −λ /n ! with expectation value λ = E • L .Each of these photons has the opportunity to get detected with a photon detection probability ξ( x), specific for interaction point x.Accordingly, the number of detected photons n, is the result of n Bernoulli trials, and stays Poisson distributed with expectation value E •L •ξ( x).Given a full event, with total coincident energy in the LAr distributed over several interaction points (E i , x i ), the probability mass function (pmf) λ s [n] for the total number of signal photons As the convolution of several independent Poisson processes, it stays a Poisson distribution described by the sum of the expectation values.Provided that ξ( x) is known, veto information can be provided on the basis of the underlying energy depositions (E i , x i ), and does not require optical simulations.It relies on the assumption that each set of photons, born from a particle's energy depositions E i , solely depends on the primary light yield L and is emitted isotropically. 2  Experiments using scintillation detectors traditionally quote the yield of detected photo-electrons (p.e.) per unit of deposited energy, i. e. the experimental light yield, in e. g. p.e./keV.This number is only meaningful, when considering a homogeneous detector, with uniform response over most of its volume.By construction this is not the case for the Gerda LAr light read-out system, whose purpose is to detect light that emerges from within and around the optically dense HPGe detector array.With this in mind, good veto performance does not necessarily go hand-in-hand with maximum experimental light yield, especially when considering background that deposits energy in the "darkest" corners of the array, where the detection probabilities are 2 These assumptions break down in the case of particle-type dependent quenching or Cherenkov radiation emission.Both effects, as well as an explicit Fano factor, could be added to Eq. 1 for future studies.Recent studies suggest the absence of further energy-dependent quenching effects in LAr [16].
primary VUV photon TPB layer WLS fiber SiPM Fig. 2: Simplified light collection chain.This onedimensional representation depicts the main material properties that affect the light collection with the Gerda fiber-SiPM instrumentation.The overall light collection efficiency for the primary VUV photon is of O(0.1) %.In real life, effects like shadowing, reflections and optical coverage enter the game.minimal, possibly even zero.Hence, it is necessary to determine the full three-dimensional map of light detection probabilities ξ( x), with special emphasis on the areas where little light is collected from.This can only be done in a dedicated Monte Carlo study, that takes into account the full photon detection chain of the Gerda LAr instrumentation.

A simple estimate
Before running such simulations, it is worthwhile to evaluate the impact of the various steps a primary VUV photon undergoes until its detection.Gerda uses a hybrid system consisting of TPB-coated WLS-fibers with SiPM-readout and PMTs to detect the LAr scintillation light that emerges from in and around the HPGe detector array.If we neglect most geometric effects, the photon detection probability can be broken down into factors that depend on basic properties of the materials and components involved.Given a primary photon that is emitted in the system LAr-TPB-fiber-SiPM, as depicted in Fig. 2, the probability ξ for its detection, can be described by Where λ is the photon wavelength.First, the VUV photon has to travel a certain distance x in LAr, while risking to get absorbed in interactions with residual impurities.The absorption length α(λ = 128 nm) at LAr peak emission is on the order of tens of centimeters, depending on the argon purity [15,17,18].The moment the VUV photon reaches and gets absorbed in any TPB layer, a blue photon with peak emission at 420 nm is re-emitted [19].The efficiency ε TPB for this process is close to 100 % [8,20].Since the typical distance for a first encounter with a TPB-coated surface is of similar order as the absorption length itself, about 1/e of the primary photons make it through this first part of the journey.Once a photon is shifted to blue, absorption in the LAr becomes negligible, as the absorption length for visible light exceeds the actual system size.Hence, it does not necessarily matter, if the blue photon directly enters a fiber at this point or later.As soon as this is the case, the photon undergoes a second wavelengthshifting step and is shifted to green with peak emission at 494 nm [21].The corresponding efficiency ε WLS , i. e. the overlap between the TPB emission and fiber absorption spectrum, is about 60 %.The green photon will stay trapped within the fiber with a trapping efficiency ε trap of about 7 % [21] and arrive at its end after about half of its absorption length of β ≈ 2 m, which adds another factor 1/ √ e.The coupling efficiency ε coupl to successfully couple the photon into the SiPM is assumed to be 80 %, whereas the photon detection efficiency (PDE) of being detected as a photo-electron signal is about 30 % at the green fiber emission [22].Multiplication of all individual contributions results in an overall detection probability of not more than 0.2 % and it can be anticipated that including geometric effects (e. g. shadowing or optical coverage) the light collection will not exceed 0.1 % for most regions of the Gerda LAr volume.

Monte Carlo implementation
The Gerda instrumentation is implemented in the Geant4-based [23][24][25] Majorana-Gerda (MaGe) simulation framework [26].For what concerns the propagation of optical photons from typical background processes, most important are the geometries enclosed by the LAr veto instrumentation as well as the optical properties of the corresponding materials.

Geometry
The HPGe detector array, including all auxiliary components, is implemented to the best available knowledge, but making reasonable approximations.The reader may find detailed technical specifications such as dimensions and materials documented in [3].The simulated setup includes: individually sized and placed HPGe detectors in their silicon/copper mounts, TPB-covered nylon mini-shrouds around each string, high-voltage and signal flat cables running from each detector to the front-end electronics, the front-end electronics themselves as well as copper structural components.Approximations are made when full degeneracy of events originating from the respective parts is expected, e. g. the level of detail of the electronics boards is low and no detailed cable routing is implemented.Details are discussed in [27].As a consequence, shadowing effects that impact the optical photon propagation, but not the standard background studies, may not be captured perfectly.
The PMTs are implemented as cylinders, with a quartz entrance window and a photo-sensitive cathode.They are placed at their respective 9 (7) positions in the top(bottom) copper plate, which to the inside is covered with a specular reflector that emulates VM2000.In contrast, the inside of the copper shrouds is lined with a diffuse PTFE reflector that represents the Tetratex ® foil.All reflector surfaces, as well as the PMT entrance windows, are covered with a wavelength-shifting TPB layer.
The fiber shroud is modeled as 15 × 6 = 90 cylinder segments covering the central part of the veto volume.Every segment contains a core, two claddings and one thin TPB layer, just as the real fibers.Their bottom ends have reflective surfaces attached, whereas optical photons reaching the upper ends are registered by photosensitive surfaces, each of them representing one SiPM.This differs from the real-world implementation, where the fibers are bent, up-routed and read-out on both ends.To avoid any misinterpretation of in-fiber correlations between channels that are connected to the same fibers, both the Monte Carlo and data signals are re-grouped to represent one channel per fiber module, resulting in a total of 9 fully independent channels.Due to sagging and uneven tensioning the optical coverage of the fibers was reduced in comparison to the maximum possible value of 75 %.In the simulation, a gap between the fiber segments parametrizes the coverage of the fiber shroud.Analyzing photos of the mounted fiber modules the real coverage was estimated to be around 50 %.Fig. 3: Emission spectra (solid lines) and absorption length (dashed lines) of indicated materials.The primary emission from the LAr follows a simple Gaussian distribution centered at 128 nm.Its absorption length connects to larger wavelength with an ad-hoc exponential scaling.Absorption and re-emission appears in TPB and the polystyrene fiber material.Nylon is only transparent to larger wavelength.

Optical properties
Fig. 3 compiles the relevant emission and absorption features implemented for the various materials.The emission of VUV scintillation photons from the LAr follows a simple Gaussian distribution centered at 128 nm with a standard deviation of 2.9 nm.It neglects contributions at longer wavelength, which have orders of magnitude lower intensity for pure LAr [28].The refractive index of the LAr is implemented using the empirical Sellmeier formalism, with the coefficients obtained in [29].Building on this, the wavelength-dependent Rayleigh scattering length is derived [30].It corresponds to about 70 cm at LAr peak emission, which is shorter than recently suggested [31].Operation of the LEGEND Liquid Argon Monitoring Apparatus (LLAMA) during the Gerda decommissioning point towards a VUV attenuation length of about 30 cm [32].Accordingly, the absorption length was set to 1/(1/30 − 1/70) ≈ 55 cm.The absorption length is modeled over the full wavelength range extending it from 128 nm with an ad-hoc exponential function.The primary VUV scintillation yield L is considered a free parameter and by default set to 28 photons/keV.A dependence of the photon yield on the incident particle, i. e. quenching, as well as characteristic singlet and triplet timing are implemented.The TPB absorption length is taken from [8], the emission spectrum from [19].Individual emission spectra, where available, are implemented for TPB on nylon [12], VM2000 [33] as well as Tetratex ® [9].The absorption length of nylon is taken from [34].Absorption and emission of the fiber material use the data presented in [21], normalized to measurements at 400 nm.The quantum efficiency of the PMTs [35] and PDE of the SiPMs [22] have been extracted from the product data sheets provided by the vendors.The reflectivities of germanium, copper, silicon and PTFE above 280 nm are taken from [36], whereas their values at VUV wavelength are largely based on assumptions.The reflectivity of VM2000 is taken from [33], the one of Tetratex ® from [37].The exact optical property values implemented in the simulation have been reported in [38].

Uncertainties
A priori , the bare simulations are not expected to reproduce the data.Details like partially inactive SiPM arrays, coating non-uniformities and shadowing by reallife cable management are not captured by the Monte Carlo implementation.Similarly, input parameters measured under conditions differing from those in Gerda, e. g. at room temperature or different wavelength, pose additional uncertainty.
Back to Eq. 1 and photon detection probabilities ξ( x): as already the number of primary VUV photons is uncertain, any linear effect, constant across the LAr volume x, is degenerate with the primary light yield L and thus only the product L • ξ( x) can be constrained by data-Monte Carlo comparison.It follows that, if a primary light yield of L = 28 photons/keV is assumed, its true value is fully absorbed in a global scaling of the efficiencies ε i , individually to each light detection channel i.The set ε i may further absorb any other global effect, e. g. an inaccurate TPB quantum efficiency, as well as any local channel-specific feature, e. g. varying photon detection efficiencies of the photo sensors.Given the large set of potential uncertainties, ε i are treated unconstrained and may take any value between zero and unity.

Parameter optimization
To obtain a predictive model of the performance of the LAr veto system, residual degrees of freedom must be removed.In the following, we describe the methodology employed to statistically infer the value of the efficiencies ε i by comparing simulated to experimental data.The full evidence of the model parameters is contained in a likelihood function, which has been maximized for special calibration data.
Given a class of events, the pmf Λ[n] that describes the number of photons n detected by some LAr veto channel is the convolution of two contributions: It is a simultaneous measurement of light from true coincidences Λ s [n] that accompany the corresponding HPGe energy deposition as well as random coincidences Λ b [n] largely produced by spectator decays such as e. g. 39 Ar in the LAr. 3 While Λ s [n] may be provided from simulations, randomly triggered events allow an evaluation of Λ b [n] from data.However, as the measured signal amplitudes suffer non-linear effects, e. g. afterpulsing and optical crosstalk, that are themselves under study and at present not implemented in the simulation, no direct pmf comparison is possible and instead the binary projection of Eq. 2 is used.In the binary "light/no-light" projection, where Λ = Λ[0] corresponds to no light, and Λ to a positive light detection, the pmf breaks down to a single expectation value, given by A positive light detection is either truly coincident without random contribution, fully random or a simultaneous detection of both.It is complementary to no detection, neither as true nor as random coincidence.Λ s (Λ b ) is the detection probability for true(random) coincidences and Λ s = Λ/Λ b quantifies the true survival probability of the underlying class of events, corrected for random coincidences.
Even though the data is reduced to binary information, the simulated pmf Λ s allows the additional detection efficiency ε to be folded into the Monte Carlo expectation.Given a count of n photons in the bare simulation, an effective detection of m < n photons can be represented as a sequence of Bernoulli trials with probability ε.The pmf Λ s [m](ε) is the result of binomial re-population throughout all n ≥ m: This technique avoids re-simulation for different values of ε. probability Λ s (ε), i. e. the chance to see one photon or more as true coincidence, is It is the inverse of no detection, i. e. the population of the "zero bin" Λ s [0](ε) in Eq. 4 and allows uncertainties on the bare frequencies Λ s [n] to be propagated into ∆Λ s (ε). 4he likelihood for the observation of light in N Monte Carlo events out of N tot simulated in total, given the aforementioned expectation value Maximizing its value allows to infer on ε, whereas taking into account the limited statistics of the random coincidence dataset, with M light detections over M tot random events, makes it a combined fit of both the data and the random coincidence sample.This combined likelihood reads: The signal expectation is given flexibility according to its uncertainty ∆Λ s using a Gaussian pull term Ĝ(σ) = e −σ 2 /2 , which accounts for limited simulation statistics and additional systematics.Eq. 6 has zero degrees of freedom and hence model discrimination can only be obtained through a combination of multiple datasets, i. e. calibration source positions, or by exploiting channel event correlations.While the former sounds trivial, e. g. an absorption length can be estimated from measurements at different distance, the latter requires explanation.Let's imagine two photosensors i ∈ {A, B}, each probing the LAr with a certain photon detection probability ξ i ( x).Considering an energy deposition (E, x), the probability to see light in both channels depends on ξ A ( x) • ξ B ( x), while events triggering only channel Both of them probe distinct regions of the LAr volume.The probability to see no light from a certain volume of the LAr, i. e. the corresponding survival probability, depends on i (1 − ξ i ( x)).
Given the full set of veto channels S of size n, each event will come as a certain subset, i. e. pattern, P ⊆ S of triggered channels.The total number of possible patterns is 2 n , where each of them comes with its own unique expectation value derived from signal as well as random coincidences.A pattern's signal expectation Λ s ( ε) can be evaluated much like Eq. 4, however starting from an n-dimensional hyper-spectrum evaluated for the full vector of efficiencies ε.When folding in the random coincidences, it has to be considered that a certain pattern P s = {A, B} may be elevated to e. g.P = {A, B, C} by random coincidences of the form P b = {A, C} or similar.Each pattern occurrence expectation value is hence a sum over all possible generator combinations G = {P s , P b }, that result in P s ∨ P b = P .The full likelihood reads The number of degrees of freedom is (2 n − n − 1), where n is the number of channels.Several datasets may be combined as the product of the individual likelihoods.Eq. 7 implicitly includes correlations between photosensors, which avoids a potential overestimation of the overall veto efficiency that could arise from unaccounted systematics.Given a large set of channels the number of possible patterns may be immense, but can be truncated by e. g. neglecting events with detection pattern higher than a certain multiplicity.

Application
In order to constrain the aforementioned effective channel efficiencies ε i , dedicated data taken under clear and reproducible conditions was needed.As the Gerda background data is a composition of various contributions and itself under study, only the peculiar conditions of a calibration run, where the energy depositions originate from a well-characterized source, allow for such studies.However, usual calibrations were taken with the purpose to guarantee a properly defined energy scale of all HPGe detectors and were performed with three 228 Th calibration sources of O( 10) kBq activity each.The resulting rate in the LAr was far too high to study the veto response and was hence not even recorded during those calibrations.For this reason, special calibration data using one of the former Phase I 228 Th sources with an activity of <2 kBq was taken in July 2016.The source was moved to 3 different vertical positions.The characteristics of the datasets are compiled in Tab. 1.To avoid β particles contributing to the coincident light production and enable a clean γ-only signature, an additional 3 mm copper housing was placed around the source container. 5Each configuration was simulated with 10 8 primary decays in the source volume.
The maximum likelihood analysis was performed on 208 Tl full energy peak (FEP) events with an energy deposit of (2615 ± 10) keV in a single HPGe detector.As no direct β transitions to the ground nor first excited state of the 208 Pb daughter nucleus are allowed, a minimum of 3.2 MeV is released in γ's, which almost always includes a transition to the intermediate 2615 keV state.Selecting full absorption events of the corresponding γ, results in an event sample virtually independent on HPGe detector details -the HPGe array is solely used to tag the 208 Tl transition.The coincident energy depositions in the LAr originate to a large extent from coincident γ's of 583 keV or more, and only marginally from Bremsstrahlung.Fig. 5 a) shows their energy distribution in the LAr.The random coincidence samples were obtained by test pulse injection at 50 mHz as well as an early (−20 μs) evaluation of the veto condition.The random coincidence appearance is unique to each source configuration as the energy depositions from independent decays in the source contribute.
To reduce the dimensionality of Eq. 7, the top/bottom PMT channels were regrouped to represent one single large top/bottom PMT, whereas the generally less uniform fiber/SiPM channels were kept separate.Accordingly, the likelihood had to be evaluated in 2 + 9 = 11 dimensions.The pattern-space was truncated so that only channel combinations present in data had to be calculated.The extracted channel efficiencies are 13 % for the top PMTs, 29 % for the bottom PMTs and reach from 21 to 37 % for the SiPM channels, with uncertainties of about ±1 %.Given an additional systematic uncertainty of 20 % on the observation of the various veto patterns (i.e. σ = 0.2 in Eq. 7), the p-value amounts to 0.2.The differences in the SiPM efficiencies match the expectation for problematic channels with potentially broken chips.The reduced value for the top PMTs was anticipated, as additional shadowing effects from cabling are only present at the top.

Probability maps
To finally evaluate the three-dimensional photon detection probability ξ( x), a dedicated simulation of VUV photons sampled uniformly over the LAr volume around the HPGe array was performed.Positive light detections with any photosensor i were determined taking into account the efficiencies ε i as obtained in Sec.5.1.For convenience, ξ( x) is stored as a discrete map, partitioned into cubic "voxels" of size 3 × 3 × 3 mm 3 .This size matches the characteristic scale of the probability map gradients expected in Gerda.Hence, the detection probability ξ k associated with voxel k corresponds to the The pdf is normalized to reproduce the total count rate in data.Despite small shape discrepancies, the predicted LAr veto survival probability (bottom panel) matches the data over a wide range of energies, even far from the model optimization energy region (gray band).A variable binning is adopted for visualization purposes.ratio between positive light detections and total scintillation photons generated in the voxel volume.Fig. 5 b) shows a projection of this object.As outlined in Sec. 3 it allows to determine the expected number of signal photons and the corresponding event rejection probability, solely based on energy depositions in the LAr.Outside the densely packed array O(0.1) %-level values are reached.Fig. 5 c) compares the energy distribution, before and after the LAr veto cut, of the 228 Th source data with the model prediction.Small discrepancies are expected from geometry inaccuracies and the modeling of charge collection at the HPGe surface.However, the event suppression (shown in the bottom panel) is reproduced over a wide range of energies far off the 208 Tl FEP-events used for model optimization.As expected, the rejection power is reduced for single-γ FEPs without significant coincidences in the decay scheme, e. g. for 212 Bi at 1621 keV, and enhanced for the double escape peak (DEP) at 1593 keV, where two 511 keV light quanta leave the HPGe detectors.

Distortion studies
As anticipated in Sec. 4, many uncertainties affect the simulation of the LAr scintillation in the Gerda setup.The channel efficiencies ε extracted from calibration data, as described in Sec. 5, absorb systematic biases that scale the detection probability by global factors (e.g. the LAr scintillation yield and the TPB quantum efficiency), but cannot cure local uncertainties (e. g. due to incorrect germanium reflectivity or fiber shroud coverage).A heuristic approach has been formulated to estimate the impact of such simulation uncertainties on the LAr veto model.The distortion of ξ( x) induced by varying input parameters can be conservatively parametrized by means of an analytical transformation T: As an additional constraint, the transformed ξ ( x) must still reproduce the calibration data presented in Sec. 5, with which the original ξ( x) was optimized.As a consequence, the LAr volumes probed by calibration data (see Fig. 5) act as a fixed point of the transformation T , letting the detection probability deviate from its nominal value in all the other regions of the setup.T may take various analytical forms, depending on the desired type of induced local distortions.The adoption of this procedure overcomes the difficulty of studying the dependence of ξ( x) on numerous optical parameters by performing several computationally expensive simulations of the scintillation light propagation.
In the context of this work, we shall focus on transformations that make ξ( x) less or more homogeneous, i. e. that make "dark" areas (e.g. the HPGe array) "darker" compared to areas with high detection probability, or vice versa.A simple transformation that meets this requirement is the following power-law scaling: where α is a real coefficient controlling the magnitude of the distortion and N is a normalization constant adjusted to reproduce the event suppression observed in calibration data.The action of the transformation in Eq. 8 on the detection probability is depicted in Fig. 6.
In the same figure, the size of power-law distortions is compared to that induced by uncertainties on the HPGe detectors' reflectivity in the VUV region.The impact of the latter has been evaluated by scaling its value by ±50 % in dedicated optical simulations.The power-law distortions significantly exceed the effect of a potential reflectivity bias and can therefore be used as a conservative estimate of the uncertainty on the light collection probability.

Background decomposition
As an application of the light collection and veto model, we shall now present the results of the background decomposition of the HPGe energy spectrum recorded during physics data taking after the application of the LAr veto cut.This background model serves as a fundamental input for various physics analyses whose sensitivity is enhanced with the LAr veto background reduction.
Previous work [27] has proven successful in describing the Gerda data in terms of background components, but before LAr veto and PSD cuts, referred to as "analysis" cuts.Probability density functions (pdfs) for various background sources have been produced with dedicated simulations of radioactive decays in the inner Gerda setup.A linear combination of these pdfs has been consequently fit on the first 60.2 kg yr data from Gerda Phase II in order to infer on the various contributions to the total background energy spectrum.Since the data was considered before the application of the LAr veto cut, the propagation of scintillation photons has been disabled in these simulations.The developed LAr detector model allows to incorporate this missing information into the existing simulations and compute background expectations after the LAr veto cut.Pdfs for a representative selection of signal and background sources in the Gerda Phase II setup are reported in Fig. 8.
The LAr veto model is applied to each background contribution before analysis cuts, as established by the background model.The distribution of α events originating from the HPGe readout contact is negligibly affected Fig. 6: Modification of the photon detection probability ξ( x) through analytical power-law distortions defined in Eq. 8. a) Inhomogeneities that are present in the nominal map are amplified with increasing α, leading to a more homogeneous (α < 1, i. e. less color contrast) or less homogeneous (α > 1, i. e. more color contrast) response.b) The probability ratio from two sample points x 1 and x 2 , outside and within the array, highlights this modification (black data points).The comparison of an altered germanium reflectivity (magenta data points) shows how the distortions conservatively exceed ±50 % on the reflectivity.by the LAr veto cut, as the α particles are expected to originate from the germanium surfaces themselves, where the light collection is poor and any remaining light output from the recoiling nucleus would be quenched [14].Thus, the α model has been imported as-is from [27].All predictions after the LAr veto cut are corrected for accidental rejection due to random coincidences of 2.7 % [4].As in [27], a Poisson likelihood is used to compare the detector-type specific energy spectra of single-detector events that survive the LAr veto cut, corresponding to an exposure of 61.4 kg yr of Gerda Phase II data 6 , with a linear combination of background pdfs.Statistical inference is carried out to determine the coefficients of the admixture that best describe the data.In a Bayesian setting, posterior probability distributions of background source intensities resulting from 6 Additional 1.2 kg yr of data from the last physics run of Phase II, neglected in previous works [27], is considered here.
the model before cuts are fed as prior information 7 , with the exception of 42 K.The distribution of 42 K ions in LAr is knowingly inhomogeneous, as ion drifts are induced by electric fields (generated by high-voltage cables and detectors) and convection.The spatial distribution is at present unknown.As a matter of fact, rough approximations have been adopted to describe it in the background model [27].Given the inhomogeneity of the LAr veto response itself, a significant mismatch of the predicted event suppression between simulation and data is expected.Hence, we adopt uninformative, uniform priors for the 42 K source intensities.The Monte Carlo Markov Chains are run with the BAT software [39] to compute posterior distributions and build knowledge update plots.Substantial agreement between event suppression predicted by the LAr veto model and data is found: the posterior distributions are compatible with the priors, where non-uniform, at the 1-2 σ level.As anticipated, the 42 K activity differs significantly from the data before the LAr veto cut.Based on this background decomposition, the event survival probability after LAr veto predicted in the 0νββ decay analysis window8 is about 0.3 % for 228 Th, 15 % for 238 U and 10 % for 60 Co.As a final remark, we stress that the quoted event suppression in the 0νββ decay region can be affected by background modeling uncertainties (e. g. source location, surface-tovolume activity or the exact 42 K spatial distribution) whose evaluation is out of the scope of this work.As such, they must be taken cum grano salis and can not be generalized for different experimental conditions.
In the energy region dominated by the two neutrino double beta (2νββ) decay, i. e. from the 39 Ar endpoint at 565 keV to the double beta (ββ) Q-value 2039 keV, excluding the intense but narrow potassium peaks, the ratio between the number of 2νββ events and the residual background influences the sensitivity of searches for ββ exotic decay modes [40].The signal-to-background ratio is about 2 in data before analysis cuts [27] and improves to about 18 after applying the LAr veto cut.
The obtained background decomposition is displayed in Fig. 7. Bands constructed by using maximally distorted maps (i.e. distortion parameter α = 0.5 and 1.5), obtained with the procedure described in Sec.6.1, are displayed for every background contribution to represent the systematic uncertainty.The ratio between data and best-fit model normalized by the expected statistical fluctuation in each bin, is shown below in the bottom panel.No significant deviations are observed, beyond the expected statistical fluctuations.
The data shown in Fig. 7 is available in ASCII format as Supplemental Material [41].

Conclusions
This paper describes the methodology, optimization and application of the light collection model as developed for the Gerda LAr scintillation light read-out.It is based on an ansatz that decouples the light propagation from nonoptical simulations, using photon detection probability maps.The model has been optimized using low-activity 228 Th calibration data.It allows predictions of the LAr veto event rejection, which is central for analyses of the 2νββ spectrum [40], including a precise determination of the 76 Ge half-life as well as a search for new physics phenomena.Even though detailed insight into the heterogeneous setup was granted, one short-coming seems imminent: certain parts of the probability maps are extrapolated, as they remain largely unprobed by the available calibration data.The systematic uncertainty associated with this problem has been evaluated using analytical distortions of the veto response.It is left to the upcoming Large Enriched Germanium Experiment for Neutrinoless double beta Decay (LEGEND) experiment to reduce this uncertainty by dedicated calibration measurements, that will elevate their LAr instrumentation from a binary light/no-light veto to a full-fledged detector.[27] (dotted lines) are overlaid for comparison.The reader is referred to [27] (Figure 1) for a detailed documentation of the simulated setup.A variable binning is adopted for visualization purposes.

Fig. 1 :
Fig. 1: LAr veto instrumentation concept.Transport of light signals towards the PMTs or SiPMs relies on wavelength-shifting processes in the TPB layers or optical fibers.Several potential light paths are indicated.Support structure details, electronics as well as individual fibers are not drawn.

Fig. 4 depictsFig. 4 :
Fig. 4: Binomial repopulation.a) The pmf Λ s [m](ε) defined in Eq. 4 can be obtained for any value of the detection efficiency ε from the unaltered simulation output.The example shows the pmf for a specific SiPM channel as obtained for 228 Th decays in a calibration source at the top of the array (see also Tab. 1) depositing (2615 ± 10) keV in the HPGe detectors.The result is plotted for a selection of efficiency values, reported in the bottom panel.b) The panel shows the non-linear dependence of the light detection probability Λ s (defined in Eq. 5).The color coding relates data points with corresponding distributions in the top panel.

Fig. 5 :
Fig. 5: Data/Monte Carlo comparison.a) Projected distribution of the LAr energy depositions for simulated 208 Tl 2615 keV FEP-events from a calibration source at position 8405 mm.Darker circles correspond to the volume occupied by the HPGe detectors.b) Photon detection probability ξ in the same region.c) Top panel: the energy spectrum of the 228 Th data corresponding to figure a), before the LAr veto cut, compared to the Monte Carlo prediction.The pdf is normalized to reproduce the total count rate in data.Despite small shape discrepancies, the predicted LAr veto survival probability (bottom panel) matches the data over a wide range of energies, even far from the model optimization energy region (gray band).A variable binning is adopted for visualization purposes.

Fig. 7 :
Fig.7: Background decomposition of the first 61.4kg yr of data from Gerda Phase II surviving the LAr veto cut (black dots).The veto model is applied to the existing background pdfs before the cut[27] folding in the probability map ξ( x).Data before the cut is shown as a light blue filled histogram.Shaded bands constructed with the maximally distorted probability maps provide a visualization of the systematic uncertainty affecting each pdf.

Fig. 8 :
Fig.8: Probability density functions (pdfs, normalized to the number of simulated primary decays) for a representative selection of background and signal event sources in the Gerda Phase II setup as detected by HPGe detectors and surviving the LAr veto cut, as predicted by the model presented in this document.Model uncertainties are shown as bands of lighter color.Pdfs before the cut[27] (dotted lines) are overlaid for comparison.The reader is referred to[27] (Figure1) for a detailed documentation of the simulated setup.A variable binning is adopted for visualization purposes.

Table 1 :
Calibration data taken with a <2 kBq228Th source placed at different heights.The reported position corresponds to the absolute distance moved from the parking position on top of the experiment.The uppermost and lower-most HPGe detectors are situated at about 8180 mm and 8560 mm respectively.