1 Introduction

Neutrinoless double beta decay is a non Standard Model process that violates the total lepton number conservation and implies a Majorana neutrino mass component [1, 2]. This decay is currently being investigated with a variety of double beta decaying isotopes. A recent review can be found in Ref. [3]. The cryogenic underground observatory for rare events (CUORE) [4,5,6] is an experiment searching for \(0\nu \beta \beta \) decay in \(^{130}\)Te. It is located at the Laboratori Nazionali del Gran Sasso of INFN, Italy. In CUORE, 988 TeO\(_2\) crystals with natural \(^{130}\)Te isotopic abundance and a 750 g average mass are operated simultaneously as source and bolometric detector for the decay. In this way, the \(0\nu \beta \beta \) decay signature is a peak at the Q-value of the reaction (\(Q_{\beta \beta }\), 2527.518 keV for \(^{130}\)Te [7,8,9]). Bolometric crystals are characterized by an excellent energy resolution (\({\sim }0.2\%\) Full Width at Half Maximum, FWHM) and a very low background at \(Q_{\beta \beta }\), which is expected to be at the \(10^{\text{- }2}\) cts\(/(\)keV\(\cdot \)kg\(\cdot \)yr\()\) level in CUORE [10].

The current best limit on \(0\nu \beta \beta \) decay in \(^{130}\)Te comes from a combined analysis of the CUORE-0 [11, 12] and Cuoricino data [13, 14]. With a total exposure of 29.6 kg\(\cdot \)year, a limit of \(T_{1/2}^{0\nu }>4.0\cdot 10^{24}\) year (\(90\%\) CI) is obtained [15] for the \(0\nu \beta \beta \) decay half life, \(T_{1/2}^{\,0\nu }\).

After the installation of the detector, successfully completed in the summer 2016, CUORE started the commissioning phase at the beginning of 2017. The knowledge of the discovery and exclusion sensitivity to \(0\nu \beta \beta \) decay as a function of the measurement live time can be exploited to set the criteria for the unblinding of the data and the release of the \(0\nu \beta \beta \) decay analysis results.

In this work, we dedicate our attention to those factors which could strongly affect the sensitivity, such as the background index (\(BI\)) and the energy resolution at \(Q_{\beta \beta }\). In CUORE, the crystals in the outer part of the array are expected to show a higher \(BI\) than those in the middle [10]. Considering this and following the strategy already implemented by the Gerda Collaboration [16, 17], we show how the division of the data into subsets with different \(BI\) could improve the sensitivity.

The reported results are obtained by means of a Bayesian analysis performed with the Bayesian analysis toolkit (BAT) [18]. The analysis is based on a toy-MC approach. At a cost of a much longer computation time with respect to the use of the median sensitivity formula [19], this provides the full sensitivity probability distribution and not only its median value.

In Sect. 2, we review the statistical methods for the parameter estimation, as well as for the extraction of the exclusion and discovery sensitivity. Section 3 describes the experimental parameters used for the analysis while its technical implementation is summarized in Sect. 4. Finally, we present the results in Sect. 5.

2 Statistical method

The computation of exclusion and discovery sensitivities presented here follows a Bayesian approach: we exploit the Bayes theorem both for parameter estimation and model comparison. In this work, we use the following notation:

  • H indicates both a hypothesis and the corresponding model;

  • \(H_0\) is the background-only hypothesis, according to which the known physics processes are enough to explain the experimental data. In the present case, we expect the CUORE background to be flat in a 100 keV region around \(Q_{\beta \beta }\), except for the presence of a \(^{60}\)Co summation peak at 2505.7 keV. Therefore, \(H_0\) is implemented as a flat background distribution plus a Gaussian describing the \(^{60}\)Co peak. In CUORE-0, this peak was found to be centered at an energy \(1.9\pm 0.7\) keV higher than that tabulated in literature [15]. This effect, present also in Cuoricino [14], is a feature of all gamma summation peaks. Hence, we will consider the \(^{60}\)Co peak to be at 2507.6 keV.

  • \(H_1\) is the background-plus-signal hypothesis, for which some new physics is required to explain the data. In our case, the physics involved in \(H_1\) contains the background processes as well as \(0\nu \beta \beta \) decay. The latter is modeled as a Gaussian peak at \(Q_{\beta \beta }\).

  • \(\mathbf {E}\) represents the data. It is a list of N energy bins centered at the energy \(E_i\) and containing \(n_i\) event counts. The energy range is [2470; 2570] keV. This is the same range used for the CUORE-0 \(0\nu \beta \beta \) decay analysis [15], and is bounded by the possible presence of peaks from \(^{214}\)Bi at 2447.7 keV and \(^{208}\)Tl X-ray escape at \({\sim }2585\) keV [15]. While an unbinned fit allows to fully exploit the information contained in the data, it can result in a long computation time for large data samples. Given an energy resolution of \({\sim }5\) keV FWHM and using a 1 keV bin width, the \(\pm 3\) sigma range of a Gaussian peak is contained in 12.7 bins. With the 1 keV binning choice, the loss of information with respect to the unbinned fit is negligible.

  • \(\Gamma ^{0\nu }\) is the parameter describing the \(0\nu \beta \beta \) decay rate for \(H_1\):

    $$\begin{aligned} \Gamma ^{0\nu } = \frac{\ln {2}}{T_{1/2}^{0\nu }}. \end{aligned}$$
    (1)
  • \(\mathbf {\theta }\) is the list of nuisance parameters describing the background processes in both \(H_0\) and \(H_1\);

  • \(\Omega \) is the parameter space for the parameters \(\mathbf {\theta }\).

2.1 Parameter estimation

We perform the parameter estimation for a model H through the Bayes theorem, which yields the probability distribution for the parameters based on the measured data, under the assumption that the model H is correct. In the \(0\nu \beta \beta \) decay analysis, we are interested in the measurement of \(\Gamma ^{0\nu }\) for the hypothesis \(H_1\). The probability distribution for the parameter set \((\Gamma ^{0\nu },\mathbf {\theta })\) is:

$$\begin{aligned}&P\left( \Gamma ^{0\nu },\mathbf {\theta } \big | \mathbf {E},H_1\right) \nonumber \\&\quad = \frac{ P\left( \mathbf {E} \big | \Gamma ^{0\nu }, \mathbf {\theta }, H_1 \right) \pi (\Gamma ^{0\nu }) \pi (\mathbf {\theta }) }{ \int _{\Omega }\int _0^{\infty } P\left( \mathbf {E} \big | \Gamma ^{0\nu }, \mathbf {\theta }, H_1 \right) \pi (\Gamma ^{0\nu }) \pi (\mathbf {\theta })\, \mathrm{d}\mathbf {\theta }\, \mathrm{d}\Gamma ^{0\nu }}. \end{aligned}$$
(2)

The numerator contains the conditional probability \(P( \mathbf {E} \big | \Gamma ^{0\nu }, \mathbf {\theta }, H_1)\) of finding the measured data \(\mathbf {E}\) given the model \(H_1\) for a set of parameters \((\Gamma ^{0\nu },\mathbf {\theta })\), times the prior probability \(\pi \) for each of the considered parameters. The prior probability has to be chosen according to the knowledge available before the analysis of the current data. For instance, the prior for the number of signal counts \(\Gamma ^{0\nu }\) might be based on the half-life limits reported by previous experiments while the prior for the background level in the region of interest (ROI) could be set based on the extrapolation of the background measured outside the ROI. The denominator represents the overall probability to obtain the data \(\mathbf {E}\) given the hypothesis \(H_1\) and all possible parameter combinations, \(P(\mathbf {E}|H_1)\).

The posterior probability distribution for \(\Gamma ^{0\nu }\) is obtained via marginalization, i.e. integrating \(P\left( \Gamma ^{0\nu },\mathbf {\theta } \big | \mathbf {E},H_1\right) \) over all nuisance parameters \(\mathbf {\theta }\):

$$\begin{aligned} P\left( \Gamma ^{0\nu } \big | H_1, \mathbf {E} \right) = \int _\Omega P\left( \Gamma ^{0\nu },\mathbf {\theta } \big | \mathbf {E},H_1\right) \mathrm{d}\mathbf {\theta }. \end{aligned}$$
(3)

For each model H, the probability of the data given the model and the parameters has to be defined. For a fixed set of experimental data, this corresponds to the likelihood function [20]. Dividing the data into \(N_d\) subsets with index d characterized by different background levels, and considering a binned energy spectrum with N bins and a number \(n_{di}\) of events in the bin i of the d subset spectrum, the likelihood function is expressed by the product of a Poisson term for each bin di:

$$\begin{aligned} P\left( \mathbf {E} \big | \Gamma ^{0\nu },\mathbf {\theta }, H \right)= & {} \mathcal {L}\left( \mathbf {E} \big | \Gamma ^{0\nu },\mathbf {\theta }, H \right) \nonumber \\= & {} \prod _{d=1}^{N_d}\ \prod _{i=1}^{N} \frac{ e^{-\lambda _{di}} \cdot \lambda _{di}^{n_{di}} }{ n_{di}!}, \end{aligned}$$
(4)

where \(\lambda _{di}\) is the expectation value for the bin di. The best-fit is defined as the set of parameter values \((\Gamma ^{0\nu },\mathbf {\theta })\) for which the likelihood is at its global maximum. In the practical case, we perform the maximization on the log-likelihood

$$\begin{aligned} \ln { \mathcal {L}\left( \mathbf {E} \big | \Gamma ^{0\nu }, \mathbf {\theta }, H \right) } = \sum _{d=1}^{N_d} \ \sum _{i=1}^{N} \left( -\lambda _{di} + \ln {\lambda _{di}^{n_{di}}} \right) , \end{aligned}$$
(5)

where the additive terms \(-\ln {(n_{di}!)}\) are dropped from the calculation.

The difference between \(H_0\) and \(H_1\) is manifested in the formulation of \(\lambda _{di}\). As mentioned above, we parametrize \(H_0\) with a flat distribution over the considered energy range, i.e. [2470; 2570] keV:

$$\begin{aligned} f_{bkg}(E) = \frac{1}{E_{\mathrm{max}} - E_{\mathrm{min}}} \end{aligned}$$
(6)

plus a Gaussian distribution for the \(^{60}\)Co peak:

$$\begin{aligned} f_{Co}(E) = \frac{1}{\sqrt{2\pi }\ \sigma } \exp {\left( - \frac{\left( E-\mu _{Co}\right) ^2}{2\sigma ^2} \right) }. \end{aligned}$$
(7)

The expected background counts in the bin di corresponds to the integral of \(f_{bkg}(E)\) in the bin di times the total number of background counts \(M^{bkg}_d\) for the subset d:

$$\begin{aligned} \lambda ^{bkg}_{di} = \int _{E_{di}^{\mathrm{min}}}^{E_{di}^{\mathrm{max}}} M^{bkg}_d f_{bkg}(E) \mathrm{d}E \end{aligned}$$
(8)

where \(E^{\mathrm{min}}_{di}\) and \(E^{\mathrm{max}}_{di}\) are the left and right margins of the energy bin di, respectively. Considering bins of size \(\delta E_{di}\) and expressing \(M^{bkg}_{di}\) as function of the background index \(BI_d\), of the total mass \(m_d\) and of the measurement live time \(t_d\), we obtain:

$$\begin{aligned} \lambda ^{bkg}_{di} = BI_d \cdot m_d \cdot t_d \cdot \delta E_{di}. \end{aligned}$$
(9)

Similarly, the expectation value for the \(^{60}\)Co distribution on the bin di is:

$$\begin{aligned} \lambda _{di}^{Co} = \int _{E^{\mathrm{min}}_{di}}^{E^{\mathrm{max}}_{di}} \frac{M^{Co}_d}{\sqrt{2\pi }\ \sigma } \exp {\left( - \frac{\left( E-\mu _{Co}\right) ^2}{2\sigma ^2} \right) }\mathrm{d}E, \end{aligned}$$
(10)

where \(M^{Co}_d\) is the total number of \(^{60}\)Co events for the subset d and can be redefined as function of the \(^{60}\)Co event rate, \(R^{Co}_d\):

$$\begin{aligned} M^{Co}_d = R^{Co}_d \cdot m_d \cdot t_d. \end{aligned}$$
(11)

The total expectation value \(\lambda _{di}\) for \(H_0\) is then:

$$\begin{aligned} \lambda _{di} = \lambda ^{bkg}_{di} + \lambda ^{Co}_{di}. \end{aligned}$$
(12)

In the case of \(H_1\) an additional expectation value for \(0\nu \beta \beta \) decay is required:

$$\begin{aligned} \lambda _{di}^{0\nu } = \int _{E^{\mathrm{min}}_{di}}^{E^{\mathrm{max}}_{di}} \frac{M^{0\nu }_d}{\sqrt{2\pi }\ \sigma } \exp {\left( - \frac{\left( E-Q_{\beta \beta }\right) ^2}{2\sigma ^2} \right) } \mathrm{d}E. \end{aligned}$$
(13)

The number of \(0\nu \beta \beta \) decay events in the subset d is:

$$\begin{aligned} M^{0\nu }_d = \Gamma ^{0\nu } \frac{ N_A }{ m_A } \cdot f_{130} \cdot \varepsilon _{\mathrm{tot}} \cdot m_d \cdot t_d, \end{aligned}$$
(14)

where \(N_A\) is the Avogadro number, \(m_a\) and \(f_{130}\) are the molar mass and the isotopic abundance of \(^{130}\)Te and \(\varepsilon _{\mathrm{tot}}\) is the total efficiency, i.e. the product of the containment efficiency \(\varepsilon _{MC}\) (obtained with MC simulations) and the instrumental efficiency \(\varepsilon _{\mathrm{instr}}\).

Fig. 1
figure 1

Top Marginalized probability distribution for \(\Gamma ^{0\nu }\) relative to one toy-MC experiment. The \(90\%\) CI limit on \(T_{1/2}^{\,0\nu }\) is indicated. Bottom distribution of the \(90\%\) CI limits on \(T^{0\nu }_{1/2}\) for \(10^5\) toy-MC experiments. The red vertical line corresponds to the median sensitivity, while the three colors depict the 1, 2 and \(3~\sigma \) quantiles of the distribution. Both plots are obtained with 1 year live time, a \(10^{\text{- }2}\) cts\(/(\)keV\(\cdot \)kg\(\cdot \)yr\()\) BI and 5 keV energy resolution

2.2 Exclusion sensitivity

We compute the exclusion sensitivity by means of the \(90\%\) CI limit. This is defined as the value of \(T_{1/2}^{\,0\nu }\) corresponding to the \(90\%\) quantile of the posterior \(\Gamma ^{0\nu }\) distribution:

$$\begin{aligned}&T_{1/2}^{\,0\nu }\left( 90\%\ CI \right) \nonumber \\&\quad = T_{1/2}^{\,0\nu }: \int _{0}^{\ln {2}/T_{1/2}^{\,0\nu }} P\left( \Gamma ^{0\nu } \big | H_1, \mathbf {E} \right) \mathrm{d}\Gamma ^{0\nu } = 0.9. \end{aligned}$$
(15)

An example of posterior probability for \(\Gamma ^{0\nu }\) and the relative \(90\%\) CI limit is shown in Fig. 1, top. Flat prior distributions are used for all parameters, as described in Sect. 3.

In the Bayesian approach, the limit is a statement regarding the true value of the considered physical quantity. In our case, a \(90\%\) CI limit on \(T_{1/2}^{\,0\nu }\) is to be interpreted as the value above which, given the current knowledge, the true value of \(T_{1/2}^{\,0\nu }\) lies with \(90\%\) probability. This differs from a frequentist \(90\%\) C.L. limit, which is a statement regarding the possible results of the repetition of identical measurements and should be interpreted as the value above which the best-fit value of \(T_{1/2}^{\,0\nu }\) would lie in the \(90\%\) of the imaginary identical experiments.

In order to extract the exclusion sensitivity, we generate a set of N toy-MC spectra according to the background-only model, \(H_0\). We then fit spectra with the background-plus-signal model, \(H_1\), and obtain the \(T_{1/2}^{\,0\nu }\left( 90\%\ CI \right) \) distribution (Fig. 1, bottom). Its median \(\hat{T}_{1/2}^{\,0\nu }\left( 90\%\ CI \right) \) is referred as the median sensitivity. For a real experiment, the experimental \(T_{1/2}^{\,0\nu }\) limit is expected to be above/below \(\hat{T}_{1/2}^{\,0\nu }\left( 90\%\ CI \right) \) with 50% probability. Alternatively, one can consider the mode of the distribution, which corresponds to the most probable \(T_{1/2}^{\,0\nu }\) limit.

The exact procedure for the computation of the exclusion sensitivity is the following:

  • for each subset, we generate a random number of background events \(N_d^{bkg}\) according to a Poisson distribution with mean \(\lambda ^{bkg}_d\);

  • for each subset, we generate \(N_d^{bkg}\) events with an energy randomly distributed according to \(f_{bkg}(E)\);

  • we repeate the procedure for the \(^{60}\)Co contribution;

  • we fit the toy-MC spectrum with the \(H_1\) model (Eq. 2), and marginalize the likelihood with respect to the parameters \(BI_d\) and \(R_d^{Co}\) (Eq. 3);

  • we extract the \(90\%\) CI limit on \(T_{1/2}^{\,0\nu }\);

  • we repeat the algorithm for N toy-MC experiments, and build the distribution of \(T_{1/2}^{\,0\nu }\left( 90\%\ CI \right) \).

2.3 Discovery sensitivity

The discovery sensitivity provides information on the required strength of the signal amplitude for claiming that the known processes alone are not sufficient to properly describe the experimental data. It is computed on the basis of the comparison between the background-only and the background-plus-signal models. A method for the calculation of the Bayesian discovery sensitivity was introduced in Ref. [21]. We report it here for completeness.

In our case, we assume that \(H_0\) and \(H_1\) are a complete set of models, for which:

$$\begin{aligned} P\bigl (H_0\big |\mathbf {E}\bigr ) + P\bigl (H_1\big |\mathbf {E}\bigr ) = 1. \end{aligned}$$
(16)

The application of the Bayes theorem to the models \(H_0\) and \(H_1\) yields:

$$\begin{aligned} P\bigl (H_0\big |\mathbf {E}\bigr )&= \frac{P\bigl ( \mathbf {E}\big |H_0 \bigr ) \pi (H_0) }{ P\bigl (\mathbf {E}\bigr )} \nonumber \\ P\bigl (H_1\big |\mathbf {E}\bigr )&= \frac{P\bigl ( \mathbf {E}\big |H_1 \bigr ) \pi (H_1) }{ P\bigl (\mathbf {E}\bigr )}. \end{aligned}$$
(17)

In this case, the numerator contains the probability of measuring the data \(\mathbf {E}\) given the model H:

$$\begin{aligned} P\bigl (\mathbf {E}\big |H_0 \bigr ) = \int _{\Omega } P\left( \mathbf {E} \big | \mathbf {\theta }, H_0 \right) \pi (\mathbf {\theta })\, \mathrm{d}\mathbf {\theta } \nonumber \\ P\bigl (\mathbf {E}\big |H_1 \bigr ) = \int _{\Omega }\int _0^{\infty } P\left( \mathbf {E} \big | \Gamma ^{0\nu }, \mathbf {\theta }, H_1 \right) \nonumber \\ \times \pi (\Gamma ^{0\nu }) \pi (\mathbf {\theta })\, \mathrm{d}\mathbf {\theta }\, \mathrm{d}\Gamma ^{0\nu }, \end{aligned}$$
(18)

while the prior probabilities for the models \(H_0\) and \(H_1\) can be chosen as 0.5 so that neither model is favored.

The denominator of Eq. 17 is the sum probability of obtaining the data \(\mathbf {E}\) given either the model \(H_0\) or \(H_1\):

$$\begin{aligned} P\bigl (\mathbf {E}\bigr ) = P\bigl ( \mathbf {E}\big |H_0 \bigr ) \pi (H_0) + P\bigl ( \mathbf {E}\big |H_1 \bigr ) \pi (H_1). \end{aligned}$$
(19)

At this point we need to define a criterion for claiming the discovery of new physics. Our choice is to quote the \(3~\sigma \) (median) discovery sensitivity, i.e. the value of \(T_{1/2}^{\,0\nu }\) for which the posterior probability of the background-only model \(H_0\) given the data is smaller than 0.0027 in 50% of the possible experiments. In other words:

$$\begin{aligned} \hat{T}_{1/2}^{0\nu }(3\sigma )= & {} T_{1/2}^{0\nu } : P\bigl ( H_0 \big | \mathbf {E} \bigr ) < 0.0027 \nonumber \\&\text {~for~}N/2~\text {experiments.} \end{aligned}$$
(20)

The detailed procedure for the determination of the discovery sensitivity is:

  • we produce a toy-MC spectrum according to the \(H_1\) model with an arbitrary value of \(T_{1/2}^{\,0\nu }\);

  • we fit the spectrum with both \(H_0\) and \(H_1\);

  • we compute \(P(H_0 | \mathbf {E})\);

  • we repeat the procedure for N toy-MC spectra using the same \(T_{1/2}^{\,0\nu }\);

  • we repeat the routine with different values of \(T_{1/2}^{\,0\nu }\) until the condition of Eq. 20 is satisfied. The iteration is implemented using the bisection method until a \(5\cdot 10^{\text{- }5}\) precision is obtained on the median \(P(H_0 | \mathbf {E})\).

3 Experimental parameters

The fit parameters of the \(H_1\) model are \(BI\), \(R^{Co}\) and \(\Gamma ^{0\nu }\), while only the first two are present for \(H_0\). If the data are divided in subsets, different \(BI\) and \(R^{Co}\) fit parameter are considered for each subset. On the contrary, the inverse \(0\nu \beta \beta \) half-life is common to all subsets.

Prior to the assembly of the CUORE crystal towers, we performed a screening survey of the employed materials [22,23,24,25,26,27,28,29]. From these measurements, either a non-zero activity was obtained, or a \(90\%\) confidence level (C.L.) upper limit was set. Additionally, the radioactive contamination of the crystals and holders was also obtained from the CUORE-0 background model [30]. We developed a full MC simulation of CUORE [10], and we used the results of the screening measurements and of the CUORE-0 background model for the normalization of the simulated spectra. We then computed the \(BI\) at \(Q_{\beta \beta }\) using the output of the simulations. In the present study, we consider only those background contributions for which a non-zero activity is obtained from the available measurements. The largest background consists of \(\alpha \) particles emitted by U and Th surface contaminations of the copper structure holding the crystals. Additionally, we consider a \(^{60}\)Co contribution normalized to the \(90\%\) C.L. limit from the screening measurement. In this sense, the effect of a \(^{60}\)Co background on the CUORE sensitivity is to be held as an upper limit. Given the \(^{60}\)Co importance especially in case of sub-optimal energy resolution, we preferred to maintain a conservative approach in this regard. In the generation of the toy-MC spectra, we take into account the \(^{60}\)Co half life (5.27 year), and set the start of data taking to January 2017.

The parameter values used for the production of the toy-MC are reported in Table 1. The quoted uncertainty on the BI comes from the CUORE MC simulations [10]. We produce the toy-MC spectra using the best-fit value of the BI. In a second time, we repeat the analysis after increasing and decreasing the BI by an amount equivalent to its statistical and systematic uncertainties combined in quadrature.

Table 1 Input parameters for the production of toy-MC spectra

After running the fit on the entire crystal array as if it were a unique detector, we considered the possibility of dividing the data grouping the crystals with a similar \(BI\). Namely, being the background at \(Q_{\beta \beta }\) dominated by surface \(\alpha \) contamination of the copper structure, the crystals facing a larger copper surface are expected to have a larger \(BI\). This effect was already observed in CUORE-0, where the crystals in the uppermost and lowermost levels, which had 3 sides facing the copper shield, were characterized by a larger background than those in all other levels, which were exposed to coppper only on 2 sides. Considering the CUORE geometry, the crystals can be divided in 4 subsets with different numbers of exposed faces. Correspondingly, they are characterized by different \(BI\), as reported in Table 2.

Table 2 Crystal subsets with different expected \(\alpha \) background in CUORE. The values of BI and \(R^{Co}\) are taken from [10]

A major ingredient of a Bayesian analysis is the choice of the priors. In the present case, we use a flat prior for all parameters. In particular, the prior distribution for \(\Gamma ^{0\nu }\) is flat between zero and a value large enough to contain \({>}99.999\%\) of its posterior distribution. This corresponds to the most conservative choice. Any other reasonable prior, e.g. a scale invariant prior on \(\Gamma ^{0\nu }\), would yield a stronger limit. A different prior choice based on the real characteristic of the experimental spectra might be more appropriate for \(BI\) and \(R^{Co}\) in the analysis of the CUORE data. For the time being the lack of data prevents the use of informative priors. As a cross-check, we performed the analysis using the \(BI\) and \(^{60}\)Co rate uncertainties obtained by the background budget as the \(\sigma \) of a Gaussian prior. No significant difference was found on the sensitivity band because the Poisson fluctuations of the generated number of background and \(^{60}\)Co events are dominant for the extraction of the \(\Gamma ^{0\nu }\) posterior probability distribution.

Table 3 lists the constant quantities present in the formulation of \(H_0\) and \(H_1\). All of them are fixed, with the exception of the live time t and the FWHM of the \(0\nu \beta \beta \) decay and \(^{60}\)Co Gaussian peaks. We perform the analysis with a FWHM of 5 and 10 keV, corresponding to a \(\sigma \) of 2.12 and 4.25 keV, respectively. Regarding the efficiency, while in the toy-MC production the BI and \(R^{Co}\) are multiplied by the instrumental efficiency,Footnote 1 in the fit the total efficiency is used. This is the product of the containment and instrumental efficiency. Also in this case, we use the same value as for CUORE-0, i.e. \(81.3\%\) [15]. We evaluate the exclusion and discovery sensitivities for different live times, with t ranging from 0.1 to 5 year and using logarithmically increasing values: \(t_{i} = 1.05\cdot t_{i-1}\).

4 Fit procedure

We perform the analysis with the software BAT v1.1.0-DEV [21], which internally uses CUBA [31] v4.2 for the integration of multi-dimensional probabilities and the Metropolis-Hastings algorithm [32] for the fit. The computation time depends on the number of samples drawn from the considered probability distribution.

For the exclusion sensitivity, we draw \(10^5\) likelihood samples for every toy-MC experiment, while, due to the higher computational cost, we use only \(10^3\) for the discovery sensitivity.

For every combination of live time, \(BI\) and energy resolution, we run \(10^5\) (\(10^3\)) toy-MC experiments for the exclusion (discovery) sensitivity study. In the case of the discovery sensitivity, we chose the number of toy-MC experiments as the minimum for which a \(2\%\) relative precision was achievable on the median sensitivity. For the exclusion sensitivity, it was possible to increase both the number of toy-MC experiments and iterations, with a systematic uncertainty on the median sensitivity at the per mil level.

Table 3 Constants used in \(H_0\) and \(H_1\)

5 Results and discussion

5.1 Exclusion sensitivity

The distributions of \(90\%\) CI limit as a function of live time with no data subdivision are shown in Fig. 2. For all \(BI\) values and all live times, the FWHM of 5 keV yields a \({\sim }45\%\) higher sensitivity with respect to a 10 keV resolution. The median sensitivity after 3 months and 5 years of data collection in the two considered cases are reported in Table 4. The dependence of the median sensitivity on live time is typical of a background-dominated experiments: namely, CUORE expects about one event every four days in a \(\pm 3\sigma \) region around \(Q_{\beta \beta }\). The results in Table 4 show also the importance of energy resolution and suggest to put a strong effort in its optimization. As a cross check, we compare the sensitivity just obtained with that provided by the analytical method presented in [19] and shown in dark green in Fig. 2. The analytical method yields a slightly higher sentitivity for short live times, while the two techniques agree when the data sample is bigger. We justify this with the fact that the uncertainty on the number of background counts obtained with the Bayesian fit is slightly larger than the corresponding Poisson uncertainty assumed in the analytical approach [33], hence the limit on \(T_{1/2}^{\,0\nu }\) is systematically weaker.Footnote 2 The effect becomes less and less strong with increasing data samples, i.e. with growing live time. With a resolution of 5 keV, the difference goes from \(8\%\) after 3 months to \({<}0.1\%\) after 5 years, while for a 10 keV FWHM the difference is \({\sim }6\%\) after 3 months and \(4\%\) after 5 years. One remark has to be done concerning the values reported in [19]: there we gave a \(90\%\) C.I. exclusion sensitivity of \(9.3\cdot 10^{25}\) year with 5 year of live time. This is \({\sim }5\%\) higher than the result presented here and is explained by the use of a different total efficiency: \(87.4\%\) in [19] and \(81.3\%\) in this work.

Fig. 2
figure 2

\(90\%\) CI exclusion sensitivity for a 5 keV (left) and a 10 keV FWHM (right). The red crosses correspond to the median sensitivity for a BI of \(10^{\text{- }2}\) cts\(/(\)keV\(\cdot \)kg\(\cdot \)yr\()\), while the smaller black crosses correspond to the median sensitivity obtained after shifting the BI up and down by an amount equivalent to its statistical and systematic uncertainty summed in quadrature. The uncertainty on BI yields a \(^{+4\%}_{-9\%}\) effect on the median sensitivity. The different colored areas depict the ranges containing the 68.3, 95.5 and \(99.7\%\) of the toy-MC experiments. They are computed for each live time value separately as described in Fig. 1, bottom. We also show the sensitivity computed as in [19] in dark green. The horizontal green line at \(4\cdot 10^{24}\) year corresponds to the limit obtained with CUORE-0 and Cuoricino [15]

We then extract the exclusion sensitivity after dividing the crystals in 4 subsets, as described in Sect. 3. The median exclusion sensitivity values after 3 months and 5 years of data collection with one and 4 subsets are reported in Table 4. The division in subsets yields only a small improvement (at the percent level) in median sensitivity. Based on this results only, one would conclude that dividing the data into subsets with different \(BI\) is not worth the effort. This conclusion is not always true, and strongly relies on the exposure and \(BI\) of the considered subsets. As an example, we repeated a toy analysis assuming a \(BI\) of \(10^{\text{- }2}\) cts\(/(\)keV\(\cdot \)kg\(\cdot \)yr\()\), and with two subsets of equal exposure and \(BI\) \(0.5\cdot 10^{\text{- }2}\) cts\(/(\)keV\(\cdot \)kg\(\cdot \)yr\()\) and \(1.5\cdot 10^{\text{- }2}\) cts\(/(\)keV\(\cdot \)kg\(\cdot \)yr\()\), respectively. In this case, the division of the data in to two subsets yields a \({\sim }10\%\) improvement after 5 year of data taking. Hence, the data subdivision is a viable option for the final analysis, whose gain strongly depends on the experimental BI of each channel. Similarly, we expect the CUORE bolometers to have different energy resolutions; in CUORE-0, these ranged from \({\sim }3\) keV to \({\sim }20\) keV FWHM [34]. In the real CUORE analysis a further splitting of the data can be done by grouping the channels with similar FWHM, or by keeping every channels separate. At the present stage it is not possible to make reliable predictions for the FWHM distribution among the crystals, so we assumed an average value (of 5 or 10 keV) throughout the whole work.

Table 4 Median exclusion sensitivity for different energy resolutions and different subset numbers

Ideally, the final CUORE \(0\nu \beta \beta \) decay analysis should be performed keeping the spectra collected by each crystal separate, additionally to the usual division of the data into data sets comprised by two calibration runs [15]. Assuming an average frequency of one calibration per month, the total number of energy spectra would be \({\sim }6\cdot 10^4\). Assuming a different but stationary \(BI\) for each crystal, and using the same \(^{60}\)Co rate for all crystals, the fit model would have \({\sim }10^3\) parameters. This represents a major obstacle for any existing implementation of the Metropolis-Hastings or Gibbs sampling algorithm. A possible way to address the problem might be the use of different algorithms, e.g. nested sampling [35, 36], or a partial analytical solution of the likelihood maximization.

We perform two further cross-checks in order to investigate the relative importance of the flat background and the \(^{60}\)Co peak. In the first scenario we set the \(BI\) to zero, and do the same for the \(^{60}\)Co rate in the second one. In both cases, the data are not divided into subsets, and resolutions of 5 and 10 keV are considered. With no flat background and a 5 keV resolution, no \(^{60}\)Co event leaks in the \(\pm 3\sigma \) region around \(Q_{\beta \beta }\) even after 5 year of measurement. As a consequence, the \(90\%\) CI limits are distributed on a very narrow band, and the median sensitivity reaches \(1.2\cdot 10^{27}\) year after 5 year of data collection. On the contrary, if we assume a 10 keV FWHM, some \(^{60}\)Co events fall in the \(0\nu \beta \beta \) decay ROI from the very beginning of the data taking. This results in a strong asymmetry of the sensitivity band. In the second cross-check, we keep the \(BI\) at \(1.02\cdot 10^{\text{- }2}\) cts\(/(\)keV\(\cdot \)kg\(\cdot \)yr\()\), but set the \(^{60}\)Co rate to zero. In both cases, the difference with respect to the standard scenario is below \(1\%\). We can conclude that the \(^{60}\)Co peak with an initial rate of 0.428 cts/(kg\(\cdot \)yr) is not worrisome for a resolution of up to 10 keV, and that the lower sensitivity obtained with 10 keV FWHM with respect to the 5 keV case is ascribable to the relative amplitude of \(\lambda ^{bkg}_{di}\) and \(\lambda ^{0\nu }_{di}\) only (Eqs. 9 and 13). This is also confirmed by the computation of the sensitivity for the optimistic scenario without the 1.9 keV shift of the \(^{60}\)Co peak used in the standard case.

We test the fit correctness and bias computing the pulls, i.e. the normalized residuals, of the number of counts assigned to each of the fit components. Denoting with \(N^{bkg}\) and \(N^{Co}\) the number of generated background and \(^{60}\)Co events, respectively, and with \(M^{bkg}\) and \(M^{Co}\) the corresponding number of reconstructed events, the pulls are defined as:

$$\begin{aligned} r_{bkg(Co)} = \frac{ M^{bkg(Co)}-N^{bkg(Co)} }{ \sigma _{M^{bkg(Co)}}}, \end{aligned}$$
(21)

where \(\sigma _{M^{bkg(Co)}}\) is the statistical uncertainty on \(M^{bkg(Co)}\) given by the fit.

For an unbiased fit, the distribution of the pulls is expected to be Gaussian with a unitary root mean square (RMS). In the case of exclusion sensitivity, we obtain \(r_{bkg}=-0.2\pm 0.4\) and \(r_{Co}=0.1\pm 0.5\) for all live times. The fact that the pull distributions are slightly shifted indicates the presence of a bias. Its origin lies in the Bayesian nature of the fit and is that all fit contributions are constrained to be greater than zero. We perform a cross-check, by extending the range of all parameters (BI, \(R^{Co}\) and \(\Gamma ^{0\nu }\)) to negative values. Under this condition, the bias disappears. In addition to this, an explanation is needed for the small RMS of the pull distributions. This is mainly due to two effects: first, the toy-MC spectra are generated using \(H_0\), while the fit is performed using \(H_1\); second, the statistical uncertainties on all parameters are larger than the Poisson uncertainty on the number of generated events. To confirm the first statement, we repeat the fit using \(H_0\) instead of \(H_1\) and we obtain pulls with zero mean and an RMS \({\sim }0.8\), which is closer to the expected value. Finally, we compare the parameter uncertainty obtained from the fit with the Poisson uncertainty for the equivalent number of counts, and we find that the difference is of \(O(20\%)\).

Fig. 3
figure 3

\(3\sigma \) discovery sensitivity with a \(BI\) of \(1.02\cdot 10^{\text{- }2}\) cts\(/(\)keV\(\cdot \)kg\(\cdot \)yr\()\) and an FWHM of 5 keV

5.2 Discovery sensitivity

The extraction of the discovery sensitivity involves fits with the background-only and the background-plus-signal models. Moreover, two multi-dimensional integrations have to be performed for each toy-MC spectrum, and a loop over the \(0\nu \beta \beta \) decay half-life has to be done until the condition of Eq. 20 is met. Due to the high computation cost, we compute the \(3~\sigma \) discovery sensitivity for a FWHM of 5 and 10 keV with no crystal subdivision. As shown in Fig. 3, with a 5 keV energy resolution CUORE has a \(3~\sigma \) discovery sensitivity superior to the limit obtained from the combined analysis of Cuore-0 and Cuoricino data [15] after less than one month of operation, and reaches \(3.7\cdot 10^{25}\) year with 5 year of live time.

Also in this case, the pulls are characterized by an RMS smaller than expected, but no bias is present due to the use of \(H_1\) for both the generation and the fit of the toy-MC spectra.

6 Conclusion and outlook

We implemented a toy-MC method for the computation of the exclusion and discovery sensitivity of CUORE using a Bayesian analysis. We have highlighted the influence of the \(BI\) and energy resolution on the exclusion sensitivity, showing how the achievement of the expected 5 keV FWHM is desirable. Additionally, we have shown how the division of the data into subsets with different \(BI\) could yield an improvement in exclusion sensitivity.

Once the CUORE data collection starts and the experimental parameters are available, the sensitivity study can be repeated in a more detailed way. As an example, non-Gaussian spectral shapes for the \(0\nu \beta \beta \) decay and \(^{60}\)Co peaks can be used, and the systematics of the energy reconstruction can be included.