Abstract
We define a new class of random probability measures, approximating the well-known normalized generalized gamma (NGG) process. Our new process is defined from the representation of NGG processes as discrete measures where the weights are obtained by normalization of the jumps of Poisson processes and the support consists of independent identically distributed location points, however considering only jumps larger than a threshold \(\varepsilon \). Therefore, the number of jumps of the new process, called \(\varepsilon \)-NGG process, is a.s. finite. A prior distribution for \(\varepsilon \) can be elicited. We assume such a process as the mixing measure in a mixture model for density and cluster estimation, and build an efficient Gibbs sampler scheme to simulate from the posterior. Finally, we discuss applications and performance of the model to two popular datasets, as well as comparison with competitor algorithms, the slice sampler and a posteriori truncation.
Similar content being viewed by others
References
Argiento, R., Guglielmi, A., Pievatolo, A.: Bayesian density estimation and model selection using nonparametric hierarchical mixtures. Comput. Stat. Data Anal. 54, 816–832 (2010)
Argiento, R., Cremaschi, A., Guglielmi, A.: A “density-based” algorithm for cluster analysis using species sampling Gaussian mixture models. J. Comput. Graph. Stat. 23, 1126–1142 (2014)
Barrios, E., Lijoi, A., Nieto-Barajas, L.E., Prünster, I.: Modeling with normalized random measure mixture models. Stat. Sci. 28, 313–334 (2013)
Bianchini, I.: A Bayesian nonparametric model for density and cluster estimation: the \(\varepsilon \)-NGG mixture model. Tesi di laurea magistrale, Ingegneria Matematica, Politecnico di Milano (2014a)
Bianchini, I.: A new finite approximation for the NGG mixture model: an application to density estimation. In: The Contribution of Young Researchers to Bayesian Statistics: Proceedings of BAYSM2014. Springer, Berlin (2015)
Caron, F.: Bayesian nonparametric models for bipartite graphs. In: NIPS, pp. 2060–2068 (2012)
Caron, F., Fox, E.B.: Bayesian nonparametric models of sparse and exchangeable random graphs. arXiv preprint. arXiv:1401.1137 (2014)
Chen, C., Ding, N., Buntine, W.: Dependent hierarchical normalized random measures for dynamic topic modeling. arXiv preprint. arXiv:1206.4671 (2012)
Cho, R.J., Campbell, M.J., Winzeler, E.A., Steinmetz, L., Conway, A., Wodicka, L., Wolfsberg, T.G., Gabrielian, A.E., Landsman, D., Lockhart, D.J., Davis, R.W.: A genome-wide transcriptional analysis of the mitotic cell cycle. Mol. Cell. 2, 65–73 (1998)
Escobar, M., West, M.: Bayesian density estimation and inference using mixtures. J. Am. Stat. Assoc. 90, 577–588 (1995)
Favaro, S., Teh, Y.: MCMC for normalized random measure mixture models. Stat. Sci. 28(3), 335–359 (2013)
Favaro, S., Walker, S.G.: Slice sampling \(\sigma \)-stable Poisson-Kingman mixture models. J. Comput. Graph. Stat. 22(4), 830–847 (2013)
Favaro, S., Guglielmi, A., Walker, S.: A class of measure-valued Markov chains and Bayesian nonparametrics. Bernoulli 18(3), 1002–1030 (2012)
Ferguson, T.S., Klass, M.: A representation of independent increment processes without Gaussian components. Ann. Math. Stat. 43, 1634–1643 (1972)
Gelfand, A.E., Kottas, A.: A computational approach for full nonparametric Bayesian inference under Dirichlet process mixture models. J. Comput. Graph. Stat. 11, 289–305 (2002)
Gradshteyn, I., Ryzhik, L.: Table of Integrals, Series, and Products, 6th edn. Academic Press, San Diego (2000)
Griffin, J., Walker, S.G.: Posterior simulation of normalized random measure mixtures. J. Comput. Graph. Stat. 20, 241–259 (2011)
Griffin, J.E.: An adaptive truncation method for inference in Bayesian nonparametric models. Stat. Comput. doi:10.1007/s11222-014-9519-4 (2014)
Griffin, J.E., Kolossiatis, M., Steel, M.F.: Comparing distributions by using dependent normalized random-measure mixtures. J. R. Stat. Soc. B 75(3), 499–529 (2013)
Ishwaran, H., James, L.: Gibbs sampling methods for stick-breaking priors. J. Am. Stat. Assoc. 96, 161–173 (2001)
Ishwaran, H., Zarepour, M.: Markov chain Monte Carlo in approximate Dirichlet and beta two-parameter process hierarchical models. Biometrika 87, 371–390 (2000)
Ishwaran, H., Zarepour, M.: Exact and approximate sum representations for the Dirichlet process. Can. J. Stat. 30, 269–283 (2002)
James, L., Lijoi, A., Prünster, I.: Posterior analysis for normalized random measures with independent increments. Scand. J. Stat. 36, 76–97 (2009)
Kallenberg, O.: Random Measures, 4th edn. Akademie, Berlin (1983)
Kalli, M., Griffin, J.E., Walker, S.G.: Slice sampling mixture models. Stat. Comput. 21(1), 93–105 (2011)
Kingman, J.F.C.: Poisson Processes, vol. 3. Oxford University Press, Oxford (1993)
Lijoi, A., Mena, R.H., Prünster, I.: Controlling the reinforcement in Bayesian nonparametric mixture models. J. R. Stat. Soc. B 69, 715–740 (2007)
Lijoi, A., Prunster, I., Walker, S.G.: Investigating nonparametric priors with Gibbs structure. Stat. Sin. 18, 1653–1668 (2008)
Lijoi, A., Nipoti, B., Prunster, I.: Bayesian inference with dependent normalized completely random measures. Bernoulli 20, 1260–1291 (2014)
MacEachern, S.N.: Computational methods for mixture of Dirichlet process models. In: Practical Nonparametric and Semiparametric Bayesian Statistics. Lecture Notes in Statistics, vol. 133, pp. 23–43. Springer, New York (1998)
Muliere, P., Tardella, L.: Approximating distributions of random functionals of Ferguson-Dirichlet priors. Can. J. Stat. 26(2), 283–297 (1998)
Neal, R.: Markov chain sampling methods for Dirichlet process mixture models. J. Comput. Graph. Stat. 9, 249–265 (2000)
Papaspiliopoulos, O., Roberts, G.O.: Retrospective Markov chain Monte Carlo methods for Dirichlet process hierarchical models. Biometrika 95, 169–186 (2008)
Pitman, J.: Some developments of the Blackwell-Macqueen urn scheme. In: Ferguson TS, Shapley LS, Macqueen JB (eds) Statistics, Probability and Game Theory: Papers in Honor of David Blackwell. IMS Lecture Notes-Monograph Series, vol. 30, pp. 245–267. Institute of Mathematical Statistics, Hayward (1996)
Pitman, J.: Poisson-Kingman partitions. In: Science and Statistics: A Festschrift for Terry Speed. IMS Lecture Notes-Monograph Series, vol. 40, pp. 1–34. Institute of Mathematical Statistics, Hayward (2003)
Pitman, J.: Combinatorial Stochastic Processes. Lecture Notes in Mathematics, vol. 1875, pp. 1–34. Springer, New York (2006)
Regazzini, E., Lijoi, A., Prünster, I.: Distributional results for means of random measures with independent increments. Ann. Stat. 31, 560–585 (2003)
Sethuraman, J.: A constructive definition of Dirichlet priors. Stat. Sinica 4(2), 639–650 (1994)
Walker, S.G.: Sampling the Dirichlet mixture model with slices. Commun. Stat. Simulat. 36, 45–54 (2007)
Acknowledgments
The authors would like to thank the referees and the Associate Editor for their valuable help in improving this manuscript. Ilaria Bianchini was partially funded by CNR-IMATI Flagship project - Factory of the future - Fab at Hospital.
Author information
Authors and Affiliations
Corresponding author
Appendices
Appendices
1.1 Appendix 1: Proof of (5)
First observe that, since \(N_\varepsilon \) has a Poisson distribution with parameter \(\varLambda _\varepsilon \), we have
Then, formula (30) in Pitman (1996) yields
where the vector \((j_1,\dots ,j_k)\) ranges over all permutations of \(k\) elements in \(\{0,\dots ,N_\varepsilon \}\). Using the gamma function identity,
we have:
Substituting the expression of \(\rho _{\varepsilon }\) in Sect. 3:
We have crossed out the indicator function in all previous lines. If we switch the finite sum and the integral, since the integrand function does not depend on the position of the clusters \(j_i\)’s, \(i=1,\dots ,k\), but only on the sizes \(n_i\), and there are \((N_\varepsilon +1)(N_\varepsilon )\dots (N_\varepsilon +1-k)={(N_\varepsilon +1)!}/(N_{\varepsilon }+1-k)!\) sequences of \(k\) distinct elements from \(\{0,\dots ,N_\varepsilon \}\), we get:
Observe that, because of the indicator function in the above formula, summation in (12) has to be taken for \(N_\varepsilon \) from \(k-1\) to \(+\infty \). Then, by the change of variable \(N_{na}=N_\varepsilon +1-k\) in the summation (\(N_\varepsilon +1-k\) is the number of non-allocated jumps), simple calculations give
By Fubini’s theorem, we can switch integration and summation, and introduce \(\varLambda _{\varepsilon ,u}\) as defined in (6), so that
that is (5), since
1.2 Appendix 2: Lemma 1
We report this simple lemma for a thorough understanding of Lemma 2.
Lemma 1
Let \((a_n)\) and \((b_n)\) be two sequences of real numbers, such that
where \(l,a_0,b_0\) are finite, and \(a_0+b_0=l\). Then
Proof
By definition of \(\liminf \) and \(\limsup \) we have:
From the hypotheses we have
so that \(\limsup b_n=b_0\), but by hypothesis \(b_0=\liminf b_n\), and consequently
We prove similarly that \(\lim _{n\rightarrow +\infty }a_{n}=a_0\).
Of course, this lemma can be generalized to any finite number of sequences.
1.3 Appendix 3: Lemma 2
Lemma 2
Let \(p_\varepsilon \) be the eppf of a \(\varepsilon -\)NGG\((\sigma ,\kappa ,\omega , P_0)\) process. Then for each \(n_1,\dots ,n_k\ \in \mathbb {N}\) with \(k\ge 0\) and \(\sum _{i=1}^{k}n_i\) \(=n\), we have
where \(p_0(\cdot )\) is the eppf of a NGG\((\sigma ,\kappa ,\omega ,P_0)\) process.
Proof
Formula (5) can be restated as
where \(f_{\varepsilon }\) denotes the integrand in the right hand side of (5) itself. In addition, the eppf of a NGG\((\sigma ,\kappa ,\omega , P_0)\) process can be written as
where
see, for instance, Lijoi et al. (2007). We first show that
This is straightforward by the following remarks:
-
1.
\(\lim _{\varepsilon \rightarrow 0}\varGamma (n_i-\sigma ,(u+\omega )\varepsilon )=\varGamma (n_i-\sigma )\), for any \(i=1,2,\ldots ,k\), by the Dominated Convergence Theorem, since \(n_i-\sigma \ge 1-\sigma >0\);
-
2.
since \(\lim _{\varepsilon \rightarrow 0}\varGamma (-\sigma ,\omega \varepsilon )=+\infty \) and
$$\begin{aligned} \varGamma (1-\sigma ,x)=-\sigma \varGamma (-\sigma ,x) + x^{-\sigma }{\mathrm{e}}^{-x} \end{aligned}$$(Gradshteyn and Ryzhik 2000), we have:
$$\begin{aligned} \begin{aligned}&\lim _{\varepsilon \rightarrow 0} \frac{\varLambda _{\varepsilon ,u}+k}{\omega ^\sigma \varGamma (-\sigma ,\omega \varepsilon )}= \frac{\kappa }{\varGamma (1-\sigma )}, \\&\lim _{\varepsilon \rightarrow 0} \left( \varLambda _{\varepsilon ,u}-\varLambda _{\varepsilon } \right) = -\kappa \frac{(\omega +u)^\sigma -\omega ^\sigma }{\sigma }. \end{aligned} \end{aligned}$$
Now let \(\fancyscript{C}=\{C_1,\dots ,C_k\}\) be a partition of \(\{ 1, \dots , n \}\) with group sizes \((n_1,\dots ,n_k)\), and let \(\Pi _n\) be the set all the possible partitions of \(\{ 1, \dots , n \}\), of any size \(k=1,\ldots ,n\). Of course, by definition of eppf,
and, in particular this holds for either \(p_\varepsilon \) and \(p_0\). Moreover, by Fatou’s Lemma we have
Suppose now that for a particular sequence \(\fancyscript{C}\in \Pi _n\), we have \({\displaystyle p_0(n_1,\dots ,n_k)< \liminf _{\varepsilon \rightarrow 0}p_{\varepsilon }(n_1,\dots ,n_k)}\). In this case
that is a contradiction. Therefore we conclude that
for all \(n_1,\ldots ,n_k\), all \(k\). Summing up, we have proved so far that:
-
\(\lim _{\varepsilon \rightarrow 0}\sum _{\fancyscript{C} \in \Pi _n}^{}p_{\varepsilon }(n_1,\dots ,n_k) =1\);
-
\(\liminf _{\varepsilon \rightarrow 0} p_\varepsilon (n_1,\dots ,n_k)=p_0(n_1,\dots ,n_k)\) for all \(\fancyscript{C}=(C_1\), \(\dots ,C_k) \in \Pi _n\);
-
\(\sum _{\fancyscript{C} \in \Pi _n}^{} p_0(n_1,\dots ,n_k)=1\).
By Lemma 1, equation (14) follows.
1.4 Appendix 4: Proof of Proposition 1
As mentioned before, \(P_\varepsilon \) is a proper species sampling model, so that \(p_{\varepsilon }\) defines a probability law on the sets of all partitions of \(\mathbb {N}_n:=\{1,\dots ,n\}\), once that we have set a positive integer \(n\). Therefore, we introduce \((N_1^\varepsilon ,\ldots ,N_k^\varepsilon )\), the sizes of the blocks (in order of appearance), of the random partition \(C_{\varepsilon ,n}\) defined by \(p_\varepsilon \), for any \(\varepsilon \ge 0\). The probability distributions of \(\{(N_1^\varepsilon ,\ldots ,N_k^\varepsilon ),\varepsilon \ge 0 \}\) are proportional to the values of \(p_\varepsilon \) (for any \(\varepsilon \ge 0\)) in (2.6) in Pitman (2006). Hence, by Lemma 2, for any \(k=1,\ldots ,n\) and any \(n\),
Here \((N_1^0,\ldots ,N_k^0)\) denotes the sizes of the blocks (in order of appearance), of the random partition \(C_{0,n}\) defined by \(p_0\), the eppf of a NGG\((\sigma ,\kappa ,\omega , P_0)\) process. By formula (2.30) in Pitman (2006), we have
where \(P_j^\varepsilon \) and \(\tilde{P}_j\) are the \(j\)-th weights of a \(\varepsilon \)-NGG and a NGG process (with parameters \((\sigma ,\kappa ,\omega , P_0)\)), respectively. Note that the sequences depending on \(n\) have only a finite number of positive weights. Recall that the weak convergence of a sequence of random probability measures is equivalent to the pointwise convergence of the Laplace transforms (see Kallenberg 1983, Theorem 4.2). Let \(f(\cdot )\) be a continuous and bounded function on \(\varTheta \). If we can invert the order of the limit operations below, then we have:
Here we have introduced notation
thus (15) proves the stated convergence, conditioning on \(\{ \tau _0\), \(\tau _1,\tau _2,\ldots \}\), which are iid from \(P_0\).
To justify the interchange of the two limits above, we must prove that the sequence \(\left\{ {\mathbb {E}}\left( {\mathrm{e}}^{-\int _{}^{}fd \mu _n^{\varepsilon } } \right) , n\ge 1\right\} \) converges uniformly. To this end, it is sufficient to show that the difference between two next terms in the sequence does not depend on \(\varepsilon \); in fact, for any \(M>0\), since
we have
where \(M \ge \sup f\). Let now \(C_{\varepsilon ,n+1}\) be a random partition on \(\{1,\dots ,n+1\}\) such that its restriction to \(\{1,\dots ,n\}\) corresponds to \(C_{\varepsilon ,n}\). We distinguish two cases:
-
1.
\(C_{\varepsilon ,n+1}\) has the same number of clusters of \(C_{\varepsilon ,n}\); one of these clusters (the \(j\)-th for instance) has one more element and, as a consequence, has size equal to \(n_j+1\);
-
2.
\(C_{\varepsilon ,n+1}\) has one more cluster than \(C_{\varepsilon ,n}\); this cluster contains only one element.
In both cases, it is not difficult to prove that
so that we are able to interchange the two limits. Finally, it is straightforward to show that the stated convergence follow from the convergence in distribution conditioning on \(\{ \tau _0,\tau _1,\tau _2,\ldots \}\), with an argument on Laplace transforms as before. This ends the first part of the Proposition, i.e. convergence for \(\varepsilon \rightarrow 0\).
Convergence as \(\varepsilon \rightarrow +\infty \) is straightforward as well. In fact, when \(\varepsilon \) increases to \(+\infty \), there are no jumps to consider in (4) but the extra \(J_0\), so that \(P_\varepsilon \) degenerates on \(\delta _{\tau _0}\).
1.5 Appendix 5: Proof of Proposition 2
The conditional distribution of \(\varvec{\theta }\) is:
where \((\theta ^*_1,\theta ^*_2,\dots , \theta ^*_k)\) is the vector of the unique values in the sample. We omit the indicator \( \mathbb {I}_{{\{1, \ldots ,N_\varepsilon +1 \}}}(k) \) till we need it. Introducing the auxiliary variable \(U\), by (13) we have:
Hence we have:
where, in this proof, \(\fancyscript{P}_0(N_\varepsilon ;\varLambda _\varepsilon )\) is the density of the Poisson distribution with parameter \(\varLambda _\varepsilon \), evaluated in \(N_\varepsilon \), and \(P_0(\tau )\) is the density of \(P_0\) evaluated in \(\tau \).
The conditional distribution of \(P_\varepsilon \), given \(U=u\) and \(\varvec{\theta }\), is as follows:
The second factor in the right handside is proportional to
Observe that, for any \(j \ne \{ l^*_1, \ldots , l^*_k \} \),
The integrand function in the second line of the formula above is the kernel of the mean intensity of a \(\varepsilon \)-NGG\((\sigma ,\kappa ,\omega +u,P_0)\) process. On the other hand, for \(i=1,\ldots ,k\):
The integrand function in (19) is the kernel of a gamma density with parameters \((n_i-\sigma ,u+\omega )\), restricted to \((\varepsilon ,+\infty )\). Summing up, we have
As in the proof of formula (5), \(N_{na}=N_\varepsilon +1-k\) is the number of non-allocated jumps. Therefore, since \(k\) is given, the conditional distribution \(\fancyscript{L}(N_\varepsilon |\) \(u,\varvec{\theta }) \) is identified by \(\fancyscript{L}(N_{na}| u,\varvec{\theta })\); we have
Let \(\varLambda _{\varepsilon ,u}\) be as in (6); it easily follows that
On the other hand, the first factor in the right handside of (17) can be computed by the following comment. Denote by \({\varvec{l}}^*=(l^*_1,\ldots ,l^*_k)\) is the vector of locations of the allocated jumps. From (16), it is clear that
The first factor in (22) refers to the unnormalized allocated process: the support is \({\varvec{\theta }}^*\), while the jumps follows independent restricted gamma densities, as clearly observed after (19). This shows point 2. of the Proposition.
On the other hand, the second factor in (22) shows that the non-allocated jumps are indeed the jumps of \(\varepsilon \)-NGG(\(\sigma \), \(\kappa \), \(\omega +u\), \(P_0\)) process, given that exactly \(N_{na}\) jumps of the process were obtained; moreover, the conditional distribution of \(N_{na}\) is described in (21). This shows point 1. of the Proposition.
Point 3 follows straightforward from (22).
Normalization of the jumps (allocated and non-allocated) gives 4.
With regard to 5., we need to integrate out \(N_\varepsilon \) in \(\fancyscript{L}(N_\varepsilon ,u,{\varvec{\theta }})\) displayed in (20). We have already made these computations in the proof of formula (5), and thus \(f_{U|\varvec{\theta }^{*}}(u|\varvec{\theta }^{*})\) is proportional to the integrand in (5).
1.6 Appendix 6: Details of the blocked Gibbs sampler
We explicitly derive every step of the Gibbs sampler in Fig. 1, starting from the joint distribution of data and parameters in (11).
-
1.
The first step is straightforward, since
$$\begin{aligned} \fancyscript{L}(u|{\varvec{X}},{\varvec{\theta }},P_\varepsilon ,\varepsilon ,\sigma ,\kappa ) \propto \fancyscript{L}(u, {\varvec{X}},{\varvec{\theta }},P_\varepsilon ,\varepsilon ,\sigma ,\kappa ). \end{aligned}$$ -
2.
Thanks to the hierarchical structure of the model, the following relation holds true:
$$\begin{aligned} \begin{aligned}&\fancyscript{L}(\varvec{\theta }|\varvec{X}, P_\varepsilon , \varepsilon , \sigma , \kappa , u)\propto \prod _{i=1}^{n}k(X_i; \theta _i)\sum _{j=0}^{N_\varepsilon }J_j\delta _{\tau _j}(\theta _i)\\&=\prod _{i=1}^{n}\sum _{j=0}^{N_\varepsilon }J_jk(X_i; \theta _i)\delta _{\tau _j}(\theta _i) =\prod _{i=1}^{n}J_jk(X_i; \tau _j), \end{aligned} \end{aligned}$$and this proves Step 2.
-
3.
As far as \(\fancyscript{L}(P_\varepsilon , \varepsilon , \sigma , \kappa |u, \varvec{\theta }, \varvec{X})\) is concerned, we have
$$\begin{aligned} \begin{aligned}&\fancyscript{L}(P_\varepsilon , \varepsilon , \sigma , \kappa |u, \varvec{\theta }, \varvec{X})= \fancyscript{L}(P_\varepsilon , \varepsilon , \sigma , \kappa |u, \varvec{\theta })\\&= \fancyscript{L}(P_\varepsilon |\varepsilon , \sigma , \kappa , u, \varvec{\theta }) \fancyscript{L}(\varepsilon , \sigma , \kappa | u, \varvec{\theta }), \end{aligned} \end{aligned}$$so that Step 3. can be split into two consecutive substeps. First we simulate from \( \fancyscript{L}(\varepsilon , \sigma , \kappa | u, \varvec{\theta })\) as follows: we integrate out \(N_\varepsilon \) (or equivalently \(N_{na}\)) from (20) and obtain
$$\begin{aligned} \begin{aligned}&\fancyscript{L}(\varepsilon , \sigma ,\kappa | u, \varvec{\theta }, \varvec{X}) \propto \sum _{N_{na}=0}^{+\infty } \fancyscript{L}(N_{na},\varepsilon , \sigma ,\kappa | u, \varvec{\theta }, \varvec{X}) \\&\quad = \dfrac{u^{n-1}}{\varGamma (n)}\left( \dfrac{\kappa }{\varGamma (1-\sigma )}\right) ^{k-1} \prod _{i=1}^{k}\biggl [\varGamma (n_i-\sigma , \varepsilon (u+\omega )) \biggr ] \\&\qquad \times \dfrac{(\omega +u)^{\sigma k-n}}{\omega ^{\sigma }\varGamma (-\sigma , \omega \varepsilon )}e^{\varLambda _{\varepsilon ,u}- \varLambda _{\varepsilon }} \left( \varLambda _{\varepsilon ,u}+k\right) \pi (\varepsilon )\pi (\sigma )\pi (\kappa ). \end{aligned} \end{aligned}$$In practical terms, Step 3a can be obtained in three substeps:
$$\begin{aligned}&\fancyscript{L}(\varepsilon | u, \varvec{\theta }, \varvec{X})\propto \prod _{i=1}^{k}\varGamma (n_i-\sigma , \varepsilon (u+\omega )) {\mathrm{e}}^{(\varLambda _{\varepsilon ,u}-\varLambda _{\varepsilon })}\nonumber \\&\quad \times \dfrac{\varLambda _{\varepsilon ,u}+k}{\varGamma (-\sigma , \omega \varepsilon )} \pi (\varepsilon ), \end{aligned}$$(23)$$\begin{aligned}&\fancyscript{L}(\sigma |u, \varvec{\theta }, \varvec{X}) \propto \dfrac{(u+\omega )^{k\sigma }}{\omega ^\sigma }\dfrac{\varLambda _{\varepsilon ,u}+k}{\varGamma (-\sigma , \omega \varepsilon )} {\mathrm{e}}^{\left( \varLambda _{\varepsilon ,u}-\varLambda _\varepsilon \right) } \nonumber \\&\quad \times \prod _{i=1}^{k} \varGamma (n_i-\sigma , \varepsilon (u+\omega ))\varGamma (1-\sigma )^{1-k}\pi (\sigma ), \end{aligned}$$(24)$$\begin{aligned}&\fancyscript{L}(\kappa | u, \varvec{\theta }, \varvec{X})= p_1 gamma(\alpha +k, R+\beta ) \nonumber \\&\quad +\,\, (1-p_1) gamma(\alpha +k-1, R+\beta ), \end{aligned}$$(25)where
$$\begin{aligned} R=\dfrac{\omega ^\sigma \varGamma (-\sigma , \varepsilon \omega )}{\varGamma (1-\sigma )}-\dfrac{(\omega +u)^\sigma \varGamma (-\sigma , \varepsilon (\omega +u))}{\varGamma (1-\sigma )} \end{aligned}$$and \(p_1\) is equal to
$$\begin{aligned} \frac{(\alpha +k-1)(u+\omega )^\sigma \varGamma (-\sigma , \varepsilon (\omega +u))}{ (\alpha \!+\!k\!-\!1)(u\!+\!\omega )^\sigma \varGamma (\!-\!\sigma , \varepsilon (\omega +u))\!+\!k(R+\beta )\varGamma (1-\sigma )}. \end{aligned}$$Here we assume that \(\pi (\kappa )\) is \(gamma(\alpha , \beta )\). Step 3.b consists in sampling from \(\fancyscript{L}(P_\varepsilon |\varepsilon , \sigma , \kappa , u, \varvec{\theta })\) and has already been described in Sect. 4.
Rights and permissions
About this article
Cite this article
Argiento, R., Bianchini, I. & Guglielmi, A. A blocked Gibbs sampler for NGG-mixture models via a priori truncation. Stat Comput 26, 641–661 (2016). https://doi.org/10.1007/s11222-015-9549-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11222-015-9549-6