Multivariate Density Estimation with Deep Neural Mixture Models

Albeit worryingly underrated in the recent literature on machine learning in general (and, on deep learning in particular), multivariate density estimation is a fundamental task in many applications, at least implicitly, and still an open issue. With a few exceptions, deep neural networks (DNNs) have seldom been applied to density estimation, mostly due to the unsupervised nature of the estimation task, and (especially) due to the need for constrained training algorithms that ended up realizing proper probabilistic models that satisfy Kolmogorov’s axioms. Moreover, in spite of the well-known improvement in terms of modeling capabilities yielded by mixture models over plain single-density statistical estimators, no proper mixtures of multivariate DNN-based component densities have been investigated so far. The paper fills this gap by extending our previous work on neural mixture densities (NMMs) to multivariate DNN mixtures. A maximum-likelihood (ML) algorithm for estimating Deep NMMs (DNMMs) is handed out, which satisfies numerically a combination of hard and soft constraints aimed at ensuring satisfaction of Kolmogorov’s axioms. The class of probability density functions that can be modeled to any degree of precision via DNMMs is formally defined. A procedure for the automatic selection of the DNMM architecture, as well as of the hyperparameters for its ML training algorithm, is presented (exploiting the probabilistic nature of the DNMM). Experimental results on univariate and multivariate data are reported on, corroborating the effectiveness of the approach and its superiority to the most popular statistical estimation techniques.


Introduction
Let us consider an unlabeled training set T = {x1, . . ., xn} of n independent random vectors (i.e., patterns) in a d-dimensional feature space, say R d .The patterns are assumed to be identically distributed according to an unknown probability density function (pdf) p(x).Density estimation consists in finding a model for p(x) based on T .This requires exploiting the statistical knowledge on p(x) implicitly underlying the very data sample T .To this aim, a suitable algorithm is applied, either rooted in statistics or in machine learning.The algorithm is expected to come up with a model of p(x) that fits the data well, in compliance with some well-defined statistical criterion (the maximum likelihood criterion [8] being possibly the most popular).The present paper proposes and investigates mixtures of multivariate component densities realized via deep neural networks (DNNs) for density estimation.It extends to multivariate mixture densities our previous work on univariate pdfs, presented as a workshop communication in [21].The reader is referred to that paper for a representative list of up-to-date applications where (implicitly or explicitly) density estimation is fundamental and still an open issue.

Intrinsic difficulties of density estimation via machine learning
As pointed out by Vapnik [27], density estimation is an intrinsically difficult problem, and it is still open nowadays.This latter fact is mostly due to the shortcomings of established statistical approaches, either parametric or non-parametric (the reader is referred to [25] for a list of the major drawbacks of the statistical techniques), and by the technical difficulties that arise from attempting to use artificial neural networks (ANNs) or machine learning for pdf estimation.Such difficulties stem from: (1) the unsupervised nature of the learning task, (2) the numerical instability problems entailed by pdfs, whose codomains may span the interval [0, +∞), and (3) the requirement for the resulting model to respect the axioms of probability.Furthermore, the use of maximum-likelihood (ML) training in ANNs tends to result in the "divergence problem", observed first in the realm of hybrid ANN/hidden Markov models [20].It consists in the progressive divergence of the value of the ANN connection weights as ML training proceeds, resulting in an unbounded growth of the integral of the pseudo-pdf computed by the ANN.The problem does not affect radial basis functions (RBF) networks whose hidden-to-output weights were constrained to be positive and to sum to one, as in the RBF/echo state machine for sequences proposed in [26], or in the RBF/graph neural network presented in [4] for the estimation of generalized random graphs.Unfortunately, the use of RBFs in the latter contexts is justified by its allowing for a proper algorithmic hybridization with models devised specifically for sequence/structure processing, but using RBFs as a stand-alone paradigm for density estimation is of neglectable practical interest, since they end up realizing plain Gaussian mixture models (GMM) estimated via ML.

Earlier approaches to ANN-based density estimation
In spite of these difficulties, several approaches to pdf estimation via ANNs are found in the literature [24].A critical survey of the literature is found in [25].Most of these approaches suffer from some limitations.First, a ML technique is presented in [11] where the "integral equals 1" requirement is satisfied numerically by dividing the output of a multilayer Perceptron (MLP) by the numerical integral of the function the MLP computes.No algorithms for computing the numerical integral over highdimensional spaces are handed out in [11].Nonetheless, this approach is related to the technique presented in this paper, insofar that ML will be exploited herein.Differently from [11], a multi-dimensional ad-hoc numeric integration method will be used in the following, jointly with hard constraints, over a mixture of DNNs.
Instead of estimating the pdf directly, techniques have been proposed that focus on the (theoretically equivalent) estimation of the corresponding cumulative distribution function (cdf) [28,10].Supervised training is used , e.g. the backpropagation algorithm (BP), relying on the empirical cdf of the data for generating the target outputs required in supervised training.After training the MLP model φ(•) of the cdf, the pdf can be retrieved by taking derivatives of φ(•).Unfortunately, there are drawbacks to these approaches, as well (see [25]).In particular, a good approximation of the cdf does not entail a good estimate of its derivative.Negative values of ∂φ(x)  ∂x may even occur, since a linear combination of logistics is not necessarily monotonically increasing.Finally, these algorithms apply naturally to univariate cases, but their application to multivariate scenarios is way trickier (to say the least).
The idea of exploiting synthetically-generated target outputs was applied to DNNbased pdf estimation in [25] and [22].The former relies on a unbiased variant of the Parzen-window (PW) density estimation technique [8] for labeling the training set for a DNN, known as Parzen neural network.Like in the traditional kn-Nearest Neighbor (kn-NN) statistical technique [8], the resulting model does not satisfy the axioms of probability.To the contrary, the algorithm for density estimation via feed-forward DNNs presented in [22] uses a modified loss function, optimized via gradient descent.Such a criterion function combines two terms: the loss between the network output and a synthetically-generated non-parametric estimate of the pdf at hand evaluated over the corresponding input pattern, and a loss between the integral of the function realized by the MLP and the unity value (that is the training "target" for the integral of the DNN).Efficient integration methods based on Markov chain Monte Carlo with importance sampling are used to compute the integral of the DNN and its derivatives w.r.t. the DNN parameters within the gradient-descent scheme used for the optimization.The asymptotic convergence of the resulting technique to the correct solution was formally proven in [23].The ideas behind such integration methods are exploited in this paper, as well.

Mixture densities
A generalization of plain pdf models stems from the adoption of mixture densities, where the unknown pdf is rather modeled in terms of a combination of any number of component densities [8].GMMs are the most popular instance of mixture densities [3].Mixture densities were originally intended as real-life extensions of the singlepdf parametric model (e.g., one Gaussian may not be capable to explain the data distribution but K Gaussian pdfs might as well be).Yet, there is much more than this behind the notion of mixture density.In fact, in mixture densities the different component pdfs are specialized to explain distinct latent phenomena (e.g., stochastic processes) that underlie the overall data generation process, each such phenomenon having different likelihood of occurrence w.r.t.others at diverse regions of the feature space.This suites particularly those situations where the statistical population under analysis is composed of several sub-populations, each having different distribution.Relevant examples are the following: 1. the statistical study of heterogeneity in meta-analysis [5], where samples drawn from disjoint populations (e.g., adults and children, male and female subjects, etc.) are collectively collected and have to be analyzed as a whole; 2. the modeling of unsupervised or partially-supervised [17] data samples in statistical pattern recognition [8], where each sub-population corresponds to a class or category; 3. the distribution of financial returns on the stock market depending on latent phenomena such as a political crisis or a war [7]; 4. the assessment of projectile accuracy in the military science of ballistics when shots at the same target come from multiple locations and/or from different munition types [18].
As it happens, the sub-populations in a mixture are unlikely to be distributed individually according to simple (e.g., Gaussian) pdfs.Consequently, parametric models (like GMMs) are generally not a very good fit.In fact, let ξ1, . . ., ξK be K disjoint states of nature (the outcomes of a discrete, latent random variable Ξ, each outcome corresponding to a specific sub-population), and let p(x|ξi) be the pdf that explains the distribution of the random observations x given the i-th state of the latent variable, for i = 1, . . ., K. At the whole population level the data will then be distributed according to the mixture p(x) = K i=1 P (ξi)p(x|ξi).Attempts to apply a GMM to model p(x) will not necessarily result in a one-to-one relationship between the Gaussian components in the GMM and the state-specific generative models p(x|ξi).In general, at the very least, more than one Gaussian component will be needed to model p(x|ξi).Although mixtures of mixture models offer increased modeling capabilities over plain mixture models to this end, they turned out to be unpopular due to the difficulties of estimation of their parameters [2] and their excessive sensitivity to local maxima of their criterion function (generally the likelihood of their parameters given the data).

The proposed DNN-based model
Aiming at overcoming the aforementioned difficulties encountered with the established approaches, and due to the unexploited potential of using mixture models, the paper proposes a plausible solution in the form of a mixture model built on deep neural networks.The model, called deep neural mixture model (DNMM), relies on a convex combination of component pdfs estimated by component-specific DNNs.The DNMM belongs to the broad family of non-parametric pdf estimation techniques, insofar that (differently form, say, GMMs) no assumptions on the form of the individual component pdfs are made [8].Due to the learning and generalization capabilities of DNNs, the DNMM can actually learn a general form for the mixture at hand, overcoming the drawbacks of the traditional non-parametric techniques, as well.A ML training algorithm is devised, satisfying (at least numerically) a combination of hard and soft constraints required in order to guarantee a proper probabilistic interpretation of the estimated model.The resulting machine can also be seen as a novel, special case of mixture of experts [29] having a specific task, a ML-based unsupervised training algorithm, and a particular probabilistic strategy for assigning credit to its individual experts.
This paper is the extended, journal version of a previous workshop communication that we presented in [21].Over that communication, the present article extends the treatment to DNNs, defines formally the modeling capabilities of DNMM, presents the cross-validated likelihood algorithm for DNMM model selection, and reports on the results of experiments conducted over complex, multi-dimensional pdfs (in [21] only experiments on univariate setups are reported and analyzed).

Overview of the paper
The paper is organized as follows.Section 2 is devoted to the ML-based learning algorithm for DNMMs.The family of pdfs that can actually be estimated to any degree of precision via DNMM is defined formally in Section 3.An automatic model selection algorithm that exploits the probabilistic nature of the DNMM along with the cross-validated likelihood criterion is presented in Section 4. The experimental evaluation is reported in Section 5, where the DNMM compares favorably with respect to established statistical techniques (parametric as well as non-parametric) in the task of estimating non-trivial mixtures densities having different number of components and diverse dimensionality of their definition domain.Finally, Section 6 draws the conclusions and outlines the current research directions.

DNMM: formal definition and training algorithm
Hereafter we rely on the notation introduced at the beginning of Section 1, such that our goal is to estimate p(x) from the information encapsulated within the training sample T .To this aim, we introduce a deep neural mixture model p(x|W ) having the following form: where W represents the whole set of parameters in the DNMM (that is c1, . . ., cK , W1, . . ., W k ).The mixing parameters ci shall satisfy ci ∈ [0, 1] for i = 1, . . ., K and where ϕi(x, Wi) is the function computed by a component-specific DNN whose set of learnable parameters is Wi.We say that this DNN realizes the i-th deep neural component of the DNMM.Henceforth, plain feed-forward DNNs are assumed (with arbitrary activation functions), but the following calculations can be adapted to a variety of alternate families of DNNs in a (mostly) straightforward manner.In order for the DNMM to satisfy Kolmogorov's axioms of probability, a constraint on ϕi(x, Wi)dx shall be imposed shortly.It goes without saying that each DNN in the DNMM has d input neurons (matching the dimensionality of the feature space) and a single output neuron (yielding the value of the estimated pdf), and it is expected to have as many hidden layers as needed (the automatic model selection procedure presented in Section 4 is suitable for architecture selection, as well).Without loss of generality for all the present intents and purposes, we assume1 that the random vectors of interest lie within a compact S ⊂ R d , such that S can be regarded as the definition domain of ϕi(x, Wi) for all i = 1, . . ., K. In so doing, numerical integration algorithms are viable to compute ϕi(x, Wi)dx, as well as the other integrals required shortly.Equation The choice of the form of the activation function fi(.) used in the output layer of the i-th DNN requires some precautions.Due to the very nature of pdfs, fi(.) is expected to have (at least in principle) a codomain defined as [0, +∞).This may be granted in several different ways.In this paper we use a logistic sigmoid with component-specific adaptive amplitude λi ∈ R + , that is fi(ai) = λi/(1 + exp(−ai)) as described in [19], where ai represents the current activation value for the output neuron of the i-th DNN in the mixture.Therefore, each neural component in the DNMM can stretch its output over any (component-specific) range [0, λi), which, instead of being predefined by the user, is learned from the data (just like any other parameter in Wi) in compliance with the nature of the corresponding component pdf 2 .
The DNMM training algorithm revolves around a learning rule for the mixture parameters W given T , such that at the end of the training process the quantity p(x|W ) results in a robust estimate of p(x).This is achieved by pursuing two purposes: (1) exploiting the information encapsulated in T to approximate the unknown pdf; (2) preventing the DNNs in the DNMM from developing spurious solutions, by enforcing the constraints S ϕi(x, Wi)dx = 1 for all i = 1, . . ., K. To this aim, a constrained algorithm is devised that builds on the stochastic gradient-ascent maximization of the point-wise likelihood p(xj|W ) of the DNMM given the current training pattern xj .The stochastic optimization step has to be applied iteratively for j = 1, . . ., n.This is achieved by means of an on-line, differentiable loss function C(.) defined as to be maximized with respect to the DNMM parameters W under the (hard) constraints that ci ∈ [0, 1] for i = 1, . . ., K and K i=1 ci = 1.The second term in the loss function is rather a "soft" constraint that enforces a unit integral of pi(x, Wi) over S for all i = 1, . . ., K, as sought, such that S p(x|W )dx ≃ 1.The hyper-parameter ρ ∈ R + balances the importance of the constraints, and may be used in real-life applications in order to tackle potential numerical issues.The learning rule ∆w for a generic parameter w in the DNMM is written as ∆w = η ∂C (.)  ∂w , where η ∈ R + is the learning rate.Different calculations are needed, depending on w being either (i) a mixing coefficient, i.e. w = c k , or (ii) a parameter3 within any of the DNN-based component densities.In case (i), we introduce K unconstrained latent variables γ1, . . ., γK , and we let for k = 1, . . ., K, where ς(x) = 1/(1 + e −x ).In so doing, hereafter each γ k can be treated as the unknown parameter to be estimated, in place of the corresponding c k .Consequently, higher-likelihood mixing parameters c k that satisfy the required constraints are implicitly yielded by the application of the learning rule.The latter can be written as: Next, let us focus on scenario (ii), i.e. let w be a parameter within one of the DNNs.Taking the partial derivative of C(W, xj ) with respect to w requires calculating the derivatives of the first and the second terms in the right-hand side of Equation (3), respectively.For notational convenience, hereafter we assume that w belongs to the generic k-th DNN.The partial derivative of the first term can be written as where we used Leibniz rule in the last step of the calculations.In passing, it is worth noting that Equation ( 6) is the mathematical statement of the very rationale behind the specific impact that the current training pattern x j has on the learning process for distinct neural components of the DNMM.In fact, the amount of parameter change ∆w is proportional to the probabilistic "credit" c k of the neural component at hand.Moreover, the quantities within brackets in Equation ( 6) depend on the value of the k-th DNN output over x j , as well as on its derivative.If, at any given time during the training process, ϕ k (.) does not change significantly in a neighborhood of x j (e.g. if x j lies in a high-likelihood plateau or, vice versa, in a close-to-zero plateau of ϕ k (.)) then the contribution of the first quantity within brackets is neglectable.Moreover, if ϕ k (x j ) ≃ 0 then the second term within brackets turns out to be neglectable, as well.To the contrary, the contribution of x j to the parameter adaptation of k-th component DNN turns out to be paramount if ϕ k (.) returns a high likelihood over x j and significant variations in its surroundings.
Next, Leibniz rule is used again in the calculation of the derivative of the second term in the right-hand side of Equation ( 3), which can be written as In order to compute the right-hand side of Equations ( 6) and ( 7) we need suitable algorithms for the computation of As regards the computation of we proceed as in traditional BP, or as in [19] in case w = λ k .As for the integrals, deterministic numerical quadrature integration techniques (e.g., Simpson's method, trapezoidal rule, etc.) are viable only if d = 1.In fact, in terms of computational burden, they cannot realistically scale up to higher dimensions (d ≥ 2).This is all the more critical in the light of the fact that S ∂ ∂w ϕ k (x, W k )dx shall be iteratively computed for each parameter of each DNN in the DNMM.Moreover, deterministic integration methods do not exploit at all the nature of the function under integration.Roughly speaking, herein the integrand is expected to be an "approximation" of the pdf (say, p k (x)) that explains the distribution of that specific sub-sample of T that is drawn from the k-th component of the mixture.To the contrary, accounting for the pdf of the data should drive the integration algorithm towards integration points that cover "interesting" regions4 of the domain of the integrand.For these reasons, we apply a component-oriented version of the approach we presented in [22].The resulting approach can be seen as an instance of Markov chain Monte Carlo [1].It is a non-deterministic, multi-dimensional integration technique which accounts for the component-specific probability distribution of the data.Let φ k (x) denote the integrand at hand (i.e., ϕ k (x, W k ) or ∂ϕ k (x,W k ) ∂w ).Monte Carlo with importance sampling [15] yields an approximation of the integral of φ k (x) over S in the form S φ k (x)dx ≃ V (S) m m ℓ=1 φ k ( ẋℓ ) where m properly sampled integration points ẋ1 , . . ., ẋm are used.Sampling of the ℓ-th integration point ẋℓ (for ℓ = 1, . . ., m) is obtained by drawing it at random from the mixture pdf p (k) u (x) defined as where u(x) is the uniform distribution over S, and α : N → (0, 1) is a decaying function of the number t of the DNMM training epochs5 for t = 1, . . ., T , such that α(1) ≃ 1.0 and α(T ) ≃ 0.0.As in [22] we let α(t) = 1/(1 + e ), where θ is a hyperparameter.Equation ( 8) is such that the importance sampling mechanism it implies accounts for the (estimated) component density pk (x|W k ) of the k-th latent subsample of the data, therefore respecting the natural distribution of such sub-population and focusing integration on the relevant integration points (i.e., the points having high component-specific likelihood).At the same time, since the estimates of this component pdfs are unfit during the early stage of the DNMM training, Equation (8) prescribes a (safer) sampling from a uniform distribution at the beginning (practically behaving like a plain Monte Carlo algorithm).As long as the robustness of the DNMM estimate increases, i.e. as t increases, sampling from pk (x|W k ) replaces progressively the sampling from u(x), ending up in purely non-uniform importance sampling.The form of α(t) is defined accordingly.Since ϕ k (x, W k ) is intrinsically non-negative, for t → T the sampling occurs substantially from |ϕ k (x, W k )| / S |ϕ k (x, W k )| dx, that is a sufficient condition for granting that the variance of the estimated integral vanishes and the corresponding error goes to zero [13].
Sampling from p (k) u (x) requires a viable technique for sampling from the k-th DNN in the DNMM.A variant of Markov chain Monte Carlo, namely the Metropolis-Hastings (M-H) algorithm [12], is exploited in this paper.M-H is robust to the fact that, during training, ϕ k (x, W k ) may not be properly normalized but it is proportional by construction to the corresponding pdf estimate (which is normalized properly, instead, by definition) [12].Due to its efficiency and ease of sampling, a multivariate logistic pdf with radial basis, having location x and scale σ, is used as the proposal pdf q(x ′ |x) required by M-H to generate a new candidate sample x ′ = (x ′ 1 , . . ., x ′ d ) from the current sample x = (x 1 , . . ., x d ).Formally, such a proposal pdf is defined as which can be easily sampled via the inverse transform sampling technique.The hyperparameters needed (i.e., the scale σ of the proposal pdf and the burn-in period for M-H) are fixed empirically as part of the overall model selection process (see Section 4).

Class of pdfs that can be modeled accurately via DNMMs
It is seen that not all pdfs can be estimated in an accurate manner via DNMMs.A couple of simple, univariate examples should be more than enough to convince the reader of this: the Dirac's Delta (which is not continuous and not bounded) and the standard exponential pdf (defined as p e (x) = 0 if x < 0, p e (x) = exp(−x) otherwise).In [25] we introduced the class of nonpaltry pdfs, that basically embrace all the "interesting" pdfs that are of practical interest (and, that can be estimated to any degree of precision by means of Parzen Neural Networks).The arguments handed out by [25] can be extended to DNMMs, as well.The formal definition of this class of function goes as follows.Let S be a compact subset of I d , where I d = [0, 1] d (the symbol S has the same practical meaning it had in the previous section).A continuous pdf p : R d → R is said to be nonpaltry over S if S p(x)dx = 1.Accordingly, we formally define the class P(S) of nonpaltry pdfs over S as the set of all the density functions that are nonpaltry over S. In passing, note that mixture densities built on nonpaltry component densities are, in turn, nonpaltry.Theorem 3.1 in [25] states that for any nonpaltry pdf p(•) over S at least one DNN exists that computes φ (ǫ) (•) s.t.φ (ǫ) (•) − p(•) ∞,S < ǫ for any ǫ ∈ R + .This (trivially) implies that at least one DNMM (with c =1) exists that computes φ Roughly speaking, we can estimate any nonpaltry pdf to any degree of precision by means of a DNMM.

Automatic likelihood-driven model selection strategy
The former section presented a modeling (or, approximation) capability of the family of pdfs realized by DNMMs, regardless of the actual convergence (during training) of a specific DNMM to the unknown pdf p(•) underlying the data T at hand.In practice, roughly speaking, the larger the cardinality of the training sample T , the closer to p(•) it tends to become (in probability) the estimate realized by the DNMM.This means that, for sufficiently large values of n, the likelihood of the DNMM given a separate validation set V (independently drawn from p(•), as well) tends to "increase6 " with n.This informal reasoning suggests the adoption of the maximum-likelihood criterion as an evaluation function that can be used throughout the DNMM model selection process.This is made viable by the very probabilistic nature of the DNMM, and it is not available to generic, non-probabilistic DNNs or mixtures of experts.
In short, an iterative model/hyperparameter search strategy may be developed as follows.To fix ideas, let us first consider the issue of fixing the number of neurons in a certain hidden layer of a DNN within the DNMM 7 .We start with an initial number of hidden neurons, and we increase it by u units (for a certain integer u, e.g.u = 1) at each of the following steps of the selection procedure.At each iteration (starting from the initialization) the DNMM is trained on T , and the likelihood of the resulting model given V is evaluated.The procedure is iterated for as long as this likelihood increases.We stop searching when, for τ consecutive iterations, the relative gain in likelihood between consecutive steps is lower than a fixed percentage ν%.The minimum description length principle is then applied [14], selecting the simplest model amongst those (explored throughout the last τ iterations) which yielded comparable values of the corresponding validation likelihood given V (where "comparable" means that their difference is below a user-defined threshold).It is worth noting that this approach relies on an instance of the so-called cross-validated likelihood criterion [16].The same incremental strategy is viable for selecting the number of layers within the DNNs in the mixture (e.g., starting from a single hidden layer and deepening the architecture by adding one more layer at each step).
The present cross-validated likelihood search strategy can be applied, in a straightforward manner, to the problem of selecting the other hyperparameters controlling the behavior of the DNMM learning process.In order to limit the computational burden of such an automatic model selection process, random search of the values of the hyperparameters (each choice characterized by the corresponding likelihood given V) is recommended instead of going through the step-by-step incremental approach we proposed for selecting the umber of neurons.

Experiments
Experiments are conducted on random samples drawn from multimodal mixtures having known form.Section 5.1 reports on the results obtained on univariate data using mixtures of 3-layer DNNs, while Section 5.2 presents the results of experiments involving multivariate data (of different dimensionalities) using deeper DNMMs.

Experiments with univariate data
In the present, univariate case, the random samples were drawn from mixtures p(x) of c Fisher-Tippett pdfs, i.e.
The mixing coefficients P 1 , . . ., P c were drawn at random from the uniform distribution over [0.1] and normalized in such a way that c i=1 P i = 1.The component densities of the Fisher-Tippett mixture are identified by their scales β i and locations µ i , for i = 1, . . ., c.The scales were drawn at random from the uniform distribution over (0.01, 0.9), and the locations were randomly and uniformly distributed over (0, 10).Estimation tasks were created using 1200 random patterns each, drawn from p(x), and a variable number c of component densities (namely c = 5, 10, 15 and 20).Each c-specific sample was split into a training set (with n = 800 patterns) and a validation set (the remaining 400 patterns).The integrated squared error (ISE) between p(x) and its estimate p(x), i.e. (p(x) − p(x)) 2 dx, is adopted in order to assess the performance of the different estimators.Simpson's method was applied to compute numerically the ISE.
In the present, illustrative setup we used DNMMs involving DNNs with a 3-layer architecture (input layer, hidden layer of 9 neurons, output layer).
Sigmoid activation functions were used in the hidden and the output layers, having layer-wise [19] adaptive λ.All the DNMM parameters underwent random initialization over zero-centered intervals, except for the values of λ (that were initialized to 1) and the mixing parameters that were initialized as c i = 1/K for i = 1, . . ., K. As in [22] we used a function α(t) having θ = 0.07, and we relied on m = 400 integration point.The latter ones were sampled at the beginning of each training epoch using a scale σ = 9 for the logistic proposal pdf in M-H.The burn-in period of the Markov chain in M-H was stretched over the first 500 states.The other hyperparameters of the DNMM were fixed via randomsearch based on the cross-validated likelihood criterion presented in Section 4. No normalization was applied to the values of the input data.Table 1 reports the results.DNMMs having different values of K (that is K = 4, 8 and 12) were tested and compared with 8-GMM, 16-GMM, and 32-GMM (initialized via k-means and refined via iterative ML [8]), k n -NN with unbiased k n = 1 √ n [8], and Parzen Window (PW) with standard width h n = 1/ √ n of its Gaussian kernels [8].The Table shows that, regardless of the model used, the corresponding ISE increase as a function of c, as expected.The DNMMs improve systematically (and, significantly) over the statistical techniques, whatever the value of of c.On average, both the 8-DNMM and the 16-DNMM yield a 46% relative ISE reduction over the PW (the latter turns out to be the most robust non-neural estimation algorithm).Welch's t-test (used in order to account for the different variances of the models) [9] returns a level of confidence > 90% on the statistical significance of the gap between the 8-(or, 12-) DNMM and the PW.Furthermore, the DNMMs turn out to be the stablest models overall, as proved by the values of the corresponding standard deviations (last column of the table).This is evidence of the fact that the estimation accuracy offered by the DNMMs is less sensitive to the complexity of the underlying Fisher-Tippett mixture (i.e., to c) than the accuracies yielded by the traditional statistical techniques.Finally, differences in terms of ISE are observed among the DNMMs depending on k.Nonetheless, differences between the 8-DNMM and the 12-DNMM turn out to be mild, and they depend on the complexity of the underlying pdf to be estimated (at least to some extent), as expected.

Experiments with multivariate data
Hereafter we use multivariate data drawn from mixtures of generalized extreme value distributions (GEV) [6] with null slope and having the following parametric form: where the GEV parameters Θ = (d, c T , µ 1 , . . ., µ cT , β 1 , . . ., β cT ) are defined as follows: 1. d is the dimensionality of the feature space as usual.Therefore, the generic real valued random vector x ∈ R d can be written as x = (x 1 , . . ., x d ); For each i = 1, . . ., d and j = 1, . . ., c, the values μij are random quantities, independently and uniformly drawn from the interval (0.1, 0.9), and the iid random values βij are drawn from the interval (0.03, 0.05).In the following, numerical integration is computed over the interval S = [0, 1.1] d .We generated different estimation tasks involving m-GEVs with an increasing number of components (C T = 4, 9, 16 and 25, respectively) and an increasing dimensionality of the feature space (d = 2, 4, 6 and 8, respectively).For each combination of such values for C T and d, a sample of as many as 1200 patterns was randomly drawn from m-GEV(x; Θ).Each such data sample was split into a training set (800 random vectors) and a validation set (the remaining 400 patterns), as we did before.The validation set was used to select the architectures and hyperparameters for the algorithms via ML-based random search (applying the model selection procedure presented in Section 4) on a d-and c T -specific basis.This cross-validated likelihood criterion was exploited also in order to fix a statistical baseline for assessing the performance of the DNMM, as follows.First, three established and popular statistical estimation techniques suitable for multivariate, multimodal pdf estimation were evaluated, namely the PW, the k n -NN, and the GMM (the latter was evaluated several times, for an increasing number of component Gaussians ranging from 4 to 32).In PW we used Gaussian kernels with bandwidth h n = h 1 / √ n, where h 1 was selected over the [10 −1 , 10 1 ] range.For the k n -NN we let k n = k 1 √ n, as usual, with k 1 selected via cross-validated ML.As for the GMMs, the k-means algorithm was used for the initialization of the parameters, and the latter were refined via iterative ML re-estimation [8].After ML selection of suitable hyperparameters, the best amongst these statistical models (in terms of likelihood on the validation set) was used for fixing the baseline (for the specific combination of d and c T at hand, then the whole procedure was repeated for the other combinations).The baseline was quantified as the ISE with respect to the true m-GEV underlying the data sample for those specific values of d and c T .Next, a DNMM was selected (again, based on the procedure presented in Section 4).Selection involved fixing the architecture of the DNNs (number of layers between 3 and 6, number of neurons per layer, type of activation functions) and the hyperparameters used for training the DNMM.The same architectures and hyperparameters were applied for all the DNNs in the DNMM at any step of the random search process.DNMMs were initialized as in the previous Section, and the same parameters used therein for the numeric integration (burn-in period, form of alpha(t), m, etc.) were used herein.Eventually, the ISE offered by the best after-training DNMM selected this way was evaluated w.r.t. the true m-GEV, and compared with the baseline.In so doing, the results (reported in table 2) can be expressed in terms of the percentage of relative ISE reduction (%) yielded by the best DNMM over the baseline 8 .Expressing the results of the experiments in this way has two significant advantages: (1) keeping the amount of values to be read and compared, and the size of the table, to a human-readable scale, and (2) conveying the sense of the gap between the DNMM and any other established statistical estimation technique.
It is seen that, in a quasi-systematic manner, the DNMM improves significantly over the statistical approaches.Only in two cases out of 16 the ISE yielded by the DNMM is just in line with (or, slightly worse than) the baseline.Overall, the DNMM offers an average 6.67 ± 3.29 relative ISE reduction over its closest competitor (that is, the baseline).The statistical significance of the improvement offered by the DNMM over the baseline (and, in turn, over any statistical technique) computed via Welch's t-test is high, namely ≥ 90%.Moreover, the values of the std.deviation (last column and last row of the table) show that the DNMM results in a substantially stable behavior overall.
We cannot help but pointing out that the gap between the DNMM and the best statistical estimates is not as wide as in the univariate setup.A possible rationale behind the reduced gap is likely to be found in the model selection procedure we applied: the random-search strategy applied within the cross-validated ML model selection algorithm over the multivariate data (much needed, in place of a more accurate grid search, in order to keep the overall computational burden to a realistic scale) turns out to be capable of finding suitable, yet quite suboptimal architectures and hyperparameters.This affects significantly the model selection in the DNMM (insofar that the latter has many hyperparameters to be fixed, each choice resulting in profoundly different models and, therefore, performances), while the issue is much less severe in the case of the statistical estimators (which basically involve selecting just a few hyperparameters, even one only in the PW and the k n -NN cases).Another reason for the reduced gap between the DNMM and the baseline lies in our inheriting from the univariate setup the hyperparameters for M-H: it is seen that the effectiveness of (say) a certain number m of integration points (which is suitable to univariate data) reduces significantly as the dimensionality of the feature space increases (in fact, many more integration points would be needed in order to obtain a representative coverage of the multidimensional space that could grant an accurate numerical computation of the integral of the multivariate pdf).Nevertheless, also in the present setup the results are evidence of the improved estimation and modeling capabilities of the DNMM even in complex multivariate density estimation tasks.

Conclusions
The paper presented the DNMM as a sound DNN-based model for multivariate pdf estimation that satisfies Kolmogorov's axioms.The DNMM overcomes all the usual shortcomings of traditional statistical techniques.In particular, its nonparametric modeling capabilities allow for realizing flexible component densities that are not constrained to have a predefined form, as it happens in parametric statistics.Moreover, the DNMM does not suffer from the limitations of nonparametric statistical estimators, as well, insofar that it is a learning machine (instead of a memory-based algorithm), which entails generalization, smooth and compact solutions, as well as reduced time-and space-complexity.Furthermore, the probabilistic nature of the DNMM allows for the exploita-tion of the cross-validated likelihood criterion in order to carry out the model selection relying on the ML of the hyperparameters given the validation data.
The experimental results show significant improvements yielded by the DNMM over the baseline of the statistical techniques on both the univariate and the multivariate data drawn from several complex pdfs.Future work revolves around the initialization procedure: non-uniform initialization of the mixing coefficients may turn out to be helpful in breaking potential ties, and initializing the individual DNN-based components via supervised learning of a subset of the components of a pre-es'¡timated reference mixture model (i.e., a GMM) may improve the random initialization of the DNNs parameters.

Table 1 :
Estimation of the Fisher-Tippett mixture p(x) (with n = 800) in terms of integrated squared error as a function of the number c of the Fisher-Tippett component densities.Best results are shown in boldface.(Legend: k-GMM and k-DNMM denote the GMM and the DNMM with k components, respectively). c

2 .
c T is the number of component GEVs in the mixture, and c T = c d where c denotes the (fixed) number of diverse modes of m-GEV(x; Θ) along each one of the dimensions in the definition domain of the pdf; 3. µ k = (µ k1 , . . ., µ kd ) and β k = (β k1 , . . ., β kd ) are the location and the scale vectors of k-th component GEV, respectively.They are defined in such a manner that the set {(µ k , β k )|k = 1, . . ., c d } embraces all possible combinations of c dimension-specific parameters (μ i1 , βi1 ), . . ., (μ ic , βic ) for i = 1, . . ., d (for a total of c T = c d combinations).

Table 2 :
Estimation of the multivariate mixture of generalized extreme value distributions m-GEV(x; Θ) (with n = 800) in terms of relative (%) ISE reduction over the baseline, as a function of the number C T of the component densities of the m-GEV(x; Θ) and of the dimensionality d of the feature space.