Skew Gaussian processes for classification

Gaussian processes (GPs) are distributions over functions, which provide a Bayesian nonparametric approach to regression and classification. In spite of their success, GPs have limited use in some applications, for example, in some cases a symmetric distribution with respect to its mean is an unreasonable model. This implies, for instance, that the mean and the median coincide, while the mean and median in an asymmetric (skewed) distribution can be different numbers. In this paper, we propose skew-Gaussian processes (SkewGPs) as a non-parametric prior over functions. A SkewGP extends the multivariate unified skew-normal distribution over finite dimensional vectors to a stochastic processes. The SkewGP class of distributions includes GPs and, therefore, SkewGPs inherit all good properties of GPs and increase their flexibility by allowing asymmetry in the probabilistic model. By exploiting the fact that SkewGP and probit likelihood are conjugate model, we derive closed form expressions for the marginal likelihood and predictive distribution of this new nonparametric classifier. We verify empirically that the proposed SkewGP classifier provides a better performance than a GP classifier based on either Laplace’s method or expectation propagation.

Therefore, GPs provide a principled, practical, and probabilistic approach to nonparametric regression and classification and they have successfully been applied to different domains (Rasmussen and Williams, 2006).
GPs have several desirable mathematical properties. The most appealing of them is that, for regression with Gaussian noise, the prior distribution is conjugate for the likelihood function. Therefore the Bayesian update step is analytic, as is computing the predictive distribution for the function behavior at unknown locations. In spite of their success, GPs have several known shortcomings.
First, the Gaussian distribution is not a "heavy-tailed" distribution, and so it is not robust to extreme outliers. Alternative to GPs have been proposed of which the most notable example is represented by the class of elliptical processes (Fang, 2018), such as Student-t processes (O'Hagan, 1991;Zhang et al., 2007), where any collection of function values has a desired elliptical distribution, with a covariance matrix built using a kernel.
Second, the Gaussian distribution is symmetric with respect to its mean. This implies, for instance, that the mean and the median coincide, while the mean and median in an asymmetric (skewed) distribution can be different numbers. This constraint limits GPs' flexibility and affects the coverage of their credible intervals (regions) -especially when considering that symmetry must hold for all components of the (latent) function and that, as for instance discussed by Kuss and Rasmussen (2005); Nickisch and Rasmussen (2008), the exact posterior of a GP classifier is skewed.
To overcome this second limitation, in this paper, we propose Skew-Gaussian processes (SkewGPs) as a non-parametric prior over functions. A SkewGP extends the multivariate Unified Skew-Normal distribution defined over finite dimensional vectors to a stochastic process, i.e. a distribution over infinite dimensional objects. A SkewGP is completely defined by a location function, a scale function and three additional parameters that depend on a latent dimension: a skewness function, a truncation vector and a covariance matrix. It is worth noting that a SkewGP reduces to a GP when the latent variables have dimension zero. Therefore, SkewGPs inherit all good properties of GPs and increase their flexibility by allowing asymmetry in the probabilistic model.
We focus on applying this new nonparametric model to a classification problem. In the case of parametric models, Durante (2018) shows that the posterior distribution of a probit likelihood and Gaussian prior has a unified skew-normal distribution. Such a novel result allowed the author to efficiently compute full posterior inferences for Bayesian logistic regression (for small datasets n ≈ 100). Moreover the author showed that the unified skew-normal distribution is a conjugate prior for the probit likelihood (without using this prior model for data analysis).
Here we extend this result to the nonparametric case, we derive a semi-analytical expression for the posterior distribution of the latent function and predictive probabilities for SkewGPs. The term semi-analytical is adopted to indicate that posterior inferences require the computation of the cumulative distribution function of a multivariate Gaussian distribution (i.e., the computation of Gaussian orthant probabilities). By using a new formulation (Gessner et al., 2019) of elliptical slice sampling (Murray et al., 2010), lin-ess, which permits efficient sampling from a linearly constrained (e.g., orthant) Gaussian domain, we show that we can compute efficiently posterior inferences for SkewGP binary classifiers. Lin-ess is a special case of elliptical slice sampling that leverages the analytic tractability of intersections of ellipses and hyperplanes to speed up the elliptical slice algorithm. In particular, this guarantees rejection-free sampling and it is therefore also highly parallelizable.
The main contributions of this paper are 1. we propose a new class of stochastic processes called Skew-Gaussian processes (SkewGP) which generalize GP models; 2. we show that a SkewGP prior process is conjugate for the probit likelihood thus deriving for the first time the posterior distribution of a GP classifier in an analytic form; 3. we derive an efficient way to learn the hyperparameters of SkewGP and compute Monte Carlo predictions using lin-ess, showing that our model has similar bottleneck computational complexity of GPs; 4. we evaluate the proposed SkewGP classifier against state-of-the-art implementations of the GP classifier which approximate the posterior with the Laplace method or with Expectation propagation; 5. we show on a small image classification dataset that a SkewGP prior can lead to better uncertainty quantification than a GP prior.

Background
The skew normal distribution is a continuous probability distribution that generalises the normal distribution to allow for non-zero skewness. The probability density function (PDF) of the univariate skew-normal distribution with location ξ ∈ R, scale σ > 0 and skew parameter α ∈ R is given by (O'Hagan and Leonard, 1976): where φ and Φ are the PDF and, respectively, Cumulative Distribution Function (CDF) of the standard univariate Normal distribution. This distribution has been generalised in several ways, see (Azzalini, 2013) for a review. In particular, Arellano and Azzalini (2006) provided a unification of the above generalizations within a single and tractable multivariate Unified Skew-Normal distribution that satisfies closure properties for marginals and conditionals and allows more flexibility due the introduction of additional parameters.

Unified Skew-Normal distribution
A vector z ∈ R p is said to have a multivariate Unified Skew-Normal distribution with latent skewness dimension s, z ∼ SUN p,s (ξ, Ω, ∆, γ, Γ ), if its probability density function (Azzalini, 2013, Ch.7) is: where φ p (z − ξ; Ω) represents the PDF of a multivariate Normal distribution with mean ξ ∈ R p and covariance Ω = D ΩΩ D Ω ∈ R p×p , withΩ being a correlation  matrix and D Ω a diagonal matrix containing the square root of the diagonal elements in Ω. The notation Φ s (a; M ) denotes the CDF of N s (0, M ) evaluated at a ∈ R s . The parameters γ ∈ R s , Γ ∈ R s×s , ∆ p×s of the SUN distribution are related to a latent variable that controls the skewness, in particular ∆ is called Skewness matrix. The PDF (1) is well-defined provided that the matrix i.e., M is positive definite. Note that when ∆ = 0, (1) reduces to φ p (z − ξ; Ω). Moreover we assume that Φ 0 (·) = 1, so that, for s = 0, (1) becomes a multivariate Normal distribution. The role of the latent dimension s can be briefly explained as follows. Consider now a random vector (2) and define y as the vector with distribution (x 1 | x 0 + γ > 0), then it can be shown (Azzalini, 2013, Ch. 7) that z = ξ +D Ω y ∼ SUN p,s (ξ, Ω, ∆, γ, Γ ). This representation will be used in Section 5 to draw samples from the distribution. Figure 1 shows the density of a univariate SUN distribution with latent dimensions s = 1 (a1) and s = 2 (a2). The effect of a higher latent dimension can be better observed in bivariate SUN densities as shown in Figure 2. The contours of the corresponding bivariate normal are dashed. We also plot the skewness directions given byΩ −1 ∆.

Skew-Gaussian process
In this section, we define a Skew-Gaussian Process (SkewGP). Consider the functions ξ : R p → R, a location function, Ω : R p × R p → R, a scale function, ∆ : R p → R s , the Skewness vector function, and γ ∈ R s , Γ ∈ R s×s . We say that a real function f : R p → R is distributed as a skew-Gaussian process with latent dimension s, if, for any sequence of n points x 1 , . . . , x n ∈ R p , the vector f (X) = [f (x 1 ), . . . , f (x n )] T is Skew-Gaussian distributed with parameters γ, Γ , location, scale and skewness matrices, respectively, given by .
The Skew-Gaussian distribution above is well defined provided that the matrix is positive definite for all X = {x 1 , . . . , x n } ⊂ R p and for all n. In this case we can write We detail how to select the parameters in Section 4, the proposition below shows that SkewGP is a proper stochastic process.
Proposition 1 The construction of a Skew-Gaussian process from a finite-dimensional distribution is well-posed.
All the proofs are in appendix.

Binary classification
Consider the training data D = {(x i , y i )} n i=1 , where x i ∈ R p and y i ∈ {0, 1}. We aim to build a nonparametric binary classifier. We first define a probabilistic model M by assuming that f ∼ SkewGP(ξ, Ω, ∆, γ, Γ ) and considering a probit model for the likelihood: where W = diag(2y 1 −1, . . . , 2y n −1). A SkewGP prior combined with a probit likelihood gives rise to a posterior SkewGP over functions, this because Skew-Gaussian distributions are conjugate priors for probit models. In the finite dimensional parametric case, this property was shown by Durante (2018), hereafter we extend it to the nonparametric one.
From Theorem 1 we can immediately derive the following result.
Corollary 1 The marginal likelihood of the observations D = {(x i , y i )} n i=1 given the probabilistic model M, that is the prior (5) and likelihood (6), is withγ,Γ defined in Theorem 1.
In classification, based on the training data D = {(x i , y i )} n i=1 , and given test inputs x * , we aim to predict the probability that y * = 1.

Corollary 2
The posterior predictive probability of y * = 1 given the test input x * ∈ R p and the training data whereγ * ,Γ * are the corresponding matrices of the posterior computed for the Note that the dummy value 1 2 inŷ does not influence the value of p(y * = 1|x * , X, y) and it was chosen only for mathematical convenience, as it allows for marginalization over f (x * ) and to derive the expression ofγ * ,Γ * similarly to the ones in Theorem 1. 1

Prior functions, parameters and hyperparameters
A SkewGP prior is completely defined by the location function ξ(x), the scale function Ω(x, x ), the latent dimension s ∈ N, the skewness vector function ∆(x, x ) ∈ R s and γ ∈ R s , Γ ∈ R s×s . As it is common for GPs, we will take the location function ξ(x) to be zero, although this need not be done. Let K(x, x ) be a positive definite covariance function (kernel) and let Ω = K(X, X) be the covariance matrix obtained by applying K(x, x ) elementwise to the training data X. In this paper, we propose the following way to define the location, scale and skewness functions of a SkewGP: with L ∈ R s×s is a diagonal matrix whose elements x ) (for stationary kernels) with K(x, x ) being the kernel function and σ 2 the variance parameter of the kernel, e.g., for the RBF kernel It can easily be proven that M > 0 and, therefore, (2) holds. We select the parameters of the kernel σ, , the locations r i of the pseudo-points and the phase diagonal matrix L by maximizing the marginal likelihood. In particular we exploit the lower bound (16) explained in Section 5.
Similarly to the inducing points in the sparse approximation of GPs (Quiñonero-Candela and Rasmussen, 2005;Snelson and Ghahramani, 2006;Titsias, 2009;Bauer et al., 2016), the points r i can be viewed as a set of s latent variables. However, their role is totally different from that of the inducing points, they allow us to locally modulate the skewness of SkewGP. Figure 3 shows latent functions (in gray, second column) drawn from a SkewGP with latent dimension 2 and the result of squashing these sample functions through the probit logistic function (first column). In all cases, we have considered ξ(x) = 0 and a RBF kernel with = 0.3 and σ 2 = 1. The values of the other parameters of the SkewGP are reported at the top of the plots in the first column. The green line is the mean function and the red dots represent the location of the s = 2 latent pseudo-points. For large positive values of γ 1 , γ 2 , SkewGP is equivalent to a GP (plots (a1)-(a2)). At the decreasing of γ i , i = 1, 2 (plots (b1)-(b2)), the mean shifts up and the mass of the distribution is concentrated on the top of the figure. By changing the phase (the sign of) L 22 (plots (c1)-(c2)), the mean and the mass of the distribution shift down at the location of the second pseudo-observation r 2 . We can magnifying this effect by decreasing both γ i (plots (d1)-(d2)). It is also possible to introduce skewness without changing the mean (plots (e1)-(e2)). In this latter case, r 1 = r 2 and the mass of the distribution is shifted up.

Computational complexity
Corollaries 1 and 2 provide two straightforward ways to compute the marginal likelihood and the predictive posterior probability however equations (12) and (13) both require the accurate computation of Φ s+n . Quasi-randomized Monte Carlo methods (Genz, 1992;Genz and Bretz, 2009;Botev, 2017) allows calculation of Φ s+n for small n (few hundreds observations). Therefore, these procedures are not in general suitable for medium and large n (apart from special cases Phinikettos and Gandy (2011) (2018)). We overcome this issue with an effective use of sampling for the predictive posterior and a mini-batch approach to marginal likelihood.
Posterior predictive distribution. In order to compute the predictive distribution we generate samples from the posterior distribution at training points and then exploit the closure properties of the SUN distribution to obtain samples at test points. The following result from Azzalini (2013) allows us to draw independent samples from the posterior in (7): where Tγ (0;Γ ) is the pdf of a multivariate Gaussian distribution truncated componentwise below −γ. Equation (15) is a consequence of the additive representation of skew-normal distributions, see (Azzalini, 2013, Ch. 7.1.2 and 5.1.3) for more details. Note that sampling U 0 can be achieved with standard methods, however using standard rejection sampling for the variable U 1 would incur in exponentially growing sampling time as the dimension increases. A commonly used sampling technique for this type of problems is Elliptical Slice Sampling (ess) (Murray et al., 2010) which is a Markov Chain Monte Carlo algorithm that performs inference in models with Gaussian priors. This method looks for acceptable samples along elliptical slices and by doing so drastically reduces the number of rejected samples. Recently, Gessner et al. (2019) proposed an extension of ess, called linear elliptical slice sampling (lin-ess), for multivariate Gaussian distribution truncated on a region defined by linear constraints. In particular, this approach analytically derives the acceptable regions on the elliptical slices used in ess and thus guarantees rejection-free sampling. This leads to a large speed up over ess, especially in high dimensions. Given posterior samples at the training points it is possible to compute the predictive posterior at a new test points x * thanks to the following result. Theorem 2 The posterior samples of the latent function computed at the test point x * can be obtained by sampling f (x * ) from: , with f (X) sampled from the posterior SUN n,s+n (ξ,Ω,∆,γ,Γ ) in Theorem 1, K is a kernel that defines the matrices Γ, ∆, Ω as in eq. (14) and where is a diagonal matrix containing the square root of the diagonal elements of the inner matrix in the r.h.s..
Observe that the computation of the predictive posterior requires the inversion of a n × n matrix (Ω 11 ), which has complexity O(n 3 ) with storage demands of O(n 2 ). This means we have similar bottleneck computational complexity of GPs. Moreover, note that sampling from SU N 1,s is extremely efficient when the latent dimension s is small (in the experiments we use s = 2).
Marginal likelihood. As discussed in the previous section, in practical application of SkewGP, the (hyper-)parameters of the scale function Ω(x, x ), of the skewness vector function ∆(x, x ) ∈ R s and the parameters γ ∈ R s , Γ ∈ R s×s have to be selected. As for GPs, we use Bayesian model selection to consistently set such parameters. This requires the maximization of the marginal likelihood with respect to these parameters and, therefore, it is essential to provide a fast and accurate way to evaluate the marginal likelihood. In this paper, we propose a simple approximation of the marginal likelihood that allows us to efficiently compute a lower bound.
Proposition 2 Consider the marginal likelihood p(D|M) in Corollary 1, then it holds where B 1 , . . . , B b denotes a partition of the training dataset into b random disjoint subsets, |B i | denotes the number of observations in the i-th element of the partition, γ B i ,Γ B i are the parameters of the posterior computed using only the subset B i of the data.
If the batch-size is low (in the experiments we have used |B i | = 30), then we can efficiently compute each term Φ s+|B i | (γ B i ;Γ B i ) by using a quasi-randomised Monte Carlo method. We can then optimize the hyper-parameter of SkewGP by maximizing the lower bound in (16).
Computational load and parallelization: To evaluate the computational load, we have generated artificial classification data using a probit likelihood model and drawing the latent function f (X) = [f (x 1 ), . . . , f (x n )], with x i ∼ N (0, 1), from a GP with zero mean and radial basis function kernel (lengthscale 0.5 and variance 2). We have then computed the full posterior latent function from Theorem 1, that is a SkewGP posterior. Figure 4 compares the CPU time for sampling 1000 instances of f (X) from the SkewGP posterior as a function of n for lin-ess vs. a standard Elliptical Slice Sampler (ess) (5000 burn in). 2 It can be observed the computational advantage of lin-ess with respect to ess. The average CPU time required to compute Φ s+|B i | (γ B i ;Γ B i ) with |B i | = 30, using the randomized lattice routine with 5000 points (Genz, 1992), is 0.5 seconds on a standard laptop. Since the above method is randomized, we use the simultaneous perturbation stochastic approximation algorithm (Spall, 1998) for optimizing the maximum lower bound (16). 3 Finally notice that, both the computation of the lower bound of the marginal likelihood and the sampling from the posterior via lin-ess are highly parallelizable. In fact, each term Φ s+|B i | (γ B i ;Γ B i ) can be computed independently as well as each sample in lin-ess can be sample independently (because lin-ess is rejection-free and, therefore, no "burn in" is necessary).

Properties of the Posterior
In the above sections, we have shown how to compute the posterior distribution of a SkewGP process when the likelihood is a probit model. The full conjugacy of the model allows us to prove that the posterior is again a SkewGP process. This section provides more details on the properties of the posterior and compares it with two approximations. For GP classification, there are two main alternative approximation schemes for finding a Gaussian approximation to the posterior: the Laplace's method and the Expectation-Propagation (EP) method, see, e.g. Rasmussen and Williams (2006) chapter 3. Figure 5 provides a one-dimensional illustration using a synthetic classification problem with 50 observations and scalar inputs taken from (Kuss and Rasmussen, 2005). Figure 5(a) shows the dataset and the predictive posterior probability for the Laplace and EP approximations. Moreover, by using a SkewGP prior with latent dimension s = 0 (that coincides with a GP prior), we have computed the exact SkewGP predictive posterior probability. Therefore, all three methods have the same prior: a GP with zero mean and RBF covariance function (the lengthscale and variance of the kernel are the same for all the three methods and have been set equal to the values that maximise the Laplace's approximation to the marginal likelihood). Figure 5(c) shows the posterior mean latent function and corresponding 95% credible intervals. It is evident that the true posterior (SkewGP) of the latent function is skew (see for instance for x ∈ [0, 2] and the slice plot in Figure  5(b)). Laplace's approximation peaks at the posterior mode, but places too much mass over positive values of f . The EP approximation aims to match the first two moments of the posterior and, therefore, usually obtains a better coverage of the posterior mass. That is why EP is usually the method of choice for approximate inference in GP classification. Figure 6 shows the true posterior and the two approximations for the same dataset, but now the lengthscale and variance of the kernel are the optimal values for the three methods. It is evident that the skewness of the posterior provides a better model fit to the data. Figure 7 shows the posteriors corresponding to a prior SkewGP process with latent-dimension s = 2. The red dot denotes the optimal location of the pseudopoint r 1 , while r 2 = 13.5 (their initial location were 5.8 and, respectively, 6). The additional degrees of freedom of the SkewGP prior process gave a much more satisfactory answer than that obtained from a GP prior model. By comparing Figures  6 and 7, it can be noticed that the skew-point allows us to locally modulate the skewness. Moreover, the additional degrees of freedom do not lead to overfitting, even with small data, as highlighted by the optimized location of r 2 (far away) that has not effect on the skewness of the posterior SkewGP.

Results
We have evaluated the proposed SkewGP classifier on a number of benchmark classification datasets and compared its classification accuracy with the accuracy of a Gaussian process classifier that uses either Laplace's method (GP-L) or Expectation Propagation (GP-EP) for approximate Bayesian inference. For GP-L and GP-EP, we have used the implementation available in GPy (GPy, since 2012).

Penn Machine Learning Benchmarks
From the Penn Machine Learning Benchmarks (Olson et al., 2017), we have selected 124 datasets (number of features up to 500). Since this pool includes nonbinary class datasets, we have defined a binary sub-problem by considering the first (class 0) and second (class 1) class. The resulting binarised subset of datasets includes datasets with number of instances between 100 and 7400. We have scaled the inputs to zero mean and unit standard deviation and used, as performance measure, the average information in bits of the predictions about the test targets (Kuss and Rasmussen, 2005): where p * i is the predicted probability for class 1. This score equals 1 bit if the true label is predicted with absolute certainty, 0 bits for random guessing and takes negative values if the prediction gives higher probability to the wrong class. We have assessed the above performance measure for the three classifiers by using 5-fold cross-validation. While we could use any kernel for GP-L, GP-EP and SkewGP, in this experiment we have chosen the RBF kernel with a lengthscale for each dimension. Figure 8 contrasts GP-L and GP-EP with SkewGP0 (SkewGP with s = 0) and SkewGP2 (SkewGP with s = 2). We selected s = 2 because we decided to use the same dimension for all datasets and, since there are several datasets where the ratio between the number of features and the number of instances is high, a latent dimension s > 2 leads to a number of parameters that exceeds the number of instances affecting the convergence of the maximization of the marginal likelihood. The proposed SkewGP2 and SkewGP0 outperform the other two models for most data sets. The average information score of SkewGP2 is 0.573 (average accuracy 0.904), SkewGP0 is 0.557 (acc. 0.882), GP-EP is 0.542 (acc. 0.859) and GP-L is 0.512 (acc. 0.863). This claim is supported by a statistical analysis. We have compared the three classifiers using the (nonparametric) Bayesian signed-rank test (Benavoli et al., 2014(Benavoli et al., , 2017. This test declares two classifiers practically equivalent when the difference of average information is less than 0.01 (1%). The interval [−0.01, 0.01] thus defines a region of practical equivalence (rope) for classifiers. The test returns the posterior probability of the vector [p(Cl 1 > Cl 2 ), p(Cl 1 ≈ Cl 2 ), p(Cl 1 < Cl 2 )] and, therefore, this posterior can be visualised in the probability simplex (Figure 9). For the comparison GP-L vs. GP-EP, as expected it can seen that GP-EP is better than GP-L. 4 Conversely, for GP-EP versus SkewGP0, 100% of the posterior mass is in the region in favor of SkewGP0, which is the region at the right bottom of the triangle. This confirms that SkewGP0 is practically significantly better than GP-L and GP-EP. The comparison SkewGP2 versus SkewGP0 shows that SkewGP2 has surely an average information score that is not worse than that of SkewGP0 and better with probability about 0.76.
The difference between GP-L and GP-EP, and SkewGP is that the posterior of SkewGP can be skewed. Therefore, we expect SkewGP to outperform GP-L and GP-EP on the datasets for which the posterior is far from Normal (e.g., highly skewed). To verify that we have computed the sample skewness statistics (SS) for each test input x * i : i )] and the expectation E[·] can be approximated using the posterior samples drawn as in Theorem 2. Note that SS(x * i ) = 0 for symmetric distributions. Figure 10(left) shows, for each of the 124 datasets, the difference between the average information score of SkewGP0 and GP-EP in the y-axis, and max x * i SS(x * i ) for SkewGP0 in the x-axis. We used a regression tree (green line) to detect structural changes in the mean of these data points. It is evident that, for large values of the maximum skewness statistics, SkewGP0 outperforms GP-EP (the average difference is positive). Figure 10(right) reports a similar plot for SkewGP2 and the difference is even more evident. This confirms that SkewGP on average outperforms GP-EP in those datasets where the posterior is skewed and has a similar performance otherwise.

Image classification
We have also considered an image classification task: Fashion MNIST dataset (each image is 28 × 28 = 784 pixels and there are 10 classes: 0 T-shirt/top, 1 Trouser, 2 Pullover, 3 Dress, 4 Coat, 5 Sandal, 6 Shirt, 7 Sneaker, 8 Bag, 9 Ankle boot). We randomly pooled 10000 images from the dataset and divided them into two sets, with 5000 cases for training and 5000 for testing. For each one of the 10 classes, we have defined a binary classification sub-problem by considering one class against all the other classes. We have compared GP-EP and SkewGP2, that is a SkewGP with latent dimension s = 2 (for the same reason outlined in the previous section). We have initialised r i by taking 2 random samples from the    Table 1: Image classification training data. We have also considered two different kernels: RBF and the Neural Network kernel (Williams, 1998). Table 2 reports the accuracy for each of the 10 binary classification sub-problems. For the RBF kernel, it can be noticed that SkewGP2 outperforms GP-EP in all sub-problems. For the NN kernel, the differences between the two models are less substantial (due to the higher performance of the NN kernel on this dataset) but in any case in favor of SkewGP2. We have also reported, for both the models, the computational time 5 (in minutes) needed to optimize the hyperparameters, to compute the posterior and to compute the predictions for all instances in the test set. This shows that SkewGP2 is also faster than GP-EP. 6 The last row reports the accuracy on the original multi-class classification problem obtained by using the one-vs-rest heuristic, with the only goal of showing that the more accurate estimate of the probability by SkewGP leads also to an increase in accuracy for one-vs-rest. A multi-class Laplace's approximation for GP classification was developed by Williams and Barber (1998) and other implementations are for instance discussed by Hernández-Lobato et al. (2011) and Chai (2012), we plan to address multi-class classification in future work.
Our goal is assessing the accuracy but also the quality of the probabilistic predictions. Figure 12, plot (a1), shows, for the RBF kernel case and for each instance in the test set of the binary sub-problem relative to class 8 vs. rest, the value of of the mean predicted probability of class rest for SkewGP2 (x-axis) and GP-EP (y-axis). Each instance is represented as a blue point. The red points highlight the instances that were misclassified by GP-EP. Figure (a2) shows the same plot, but the red points are now the instances that were misclassified by SkewGP2. By comparing (a1) and (a2), it is evident that SkewGP2 provides a higher quality of the probabilistic predictions.
SkewGP2 also returns a better estimate of its own uncertainty. This is showed in plots (b1) vs. (b2). For each instance in the test set and for each sample from the posterior, we have computed the predicted class (the class that has probability greater than 0.5). For each test instance, we have then computed the standard deviation of all these predictions and used it to color the scatter plot of the mean predicted probability. In this way, we have a visual representation of first order (mean predicted probability) and second order (standard deviation of the predictions) uncertainty. Plot 12(b1) is relative to GP-EP and shows that GP-EP confidence is low only for the instances whose mean predicted probability is close to 0.5. This is not reflected in the value of the mean predicted probability for the misclassified instances (compare plot (a1) and (b1)). We have also computed the histogram of the standard deviation of the predictions for those instances that were misclassified by GP-EP in Figure (c1). Note that, the peak of the histogram corresponds to very low standard deviation, that means GP-EP has misclassified instances that have low second order uncertainty. This implies that the model is overestimating its confidence. Conversely, the second order uncertainty of SkewGP2 is clearly consistent, see plot (a2) and (b2), and in particular the histogram in (c2) -the peak is in correspondence of high values of the standard deviation of the predictions. In other words, SkewGP2 has mainly misclassified instances with high second order uncertainty, that is what we expect from a calibrated probabilistic model. We have reported additional examples of the better calibration of SkewGP2 for the MNIST and German-road sign dataset in appendix.

Conclusions
We have introduced the Skew Gaussian process (SkewGP) as an alternative to Gaussian processes for classification. We have shown that SkewGP and the probit likelihood are conjugate and provided marginals and closed form conditionals. We have also shown that SkewGP contains the GP as a special case and, therefore, SkewGPs inherit all good properties of GPs and increase their flexibility. The SkewGP prior was applied in classification showing improved performance over GPs (Laplace's method and Expectation Propagation approximations).
As future work, we plan to study other more native ways to parametrize the skewness matrix ∆ that do not rely on an underlying kernel. Moreover, we plan to investigate the possibility of using inducing points, as for sparse GPs, to reduce the computational load for matrix operations (complexity O(n 3 ) with storage demands of O(n 2 )) as well as deriving the posterior for the multi-class classification problem.  Suppose that a family F I of probability measures are the I-finite-dimensional marginals of an infinite-dimensional measure F (a "stochastic process"). Each measure F I belongs to the finite-dimensional subspace of dimension I. Given two marginals F I , F J , as marginals of the same measure F , they must be marginals of each other, that is where · ↓ I denotes the projection onto the subspace of dimension I. A family of probability measures that satisfies (17) is called a projective family. The Kolmogorov's extension theorem states that any projective family on the finite-dimensional subspaces of an infinite-dimensional product space uniquely defines a stochastic process on the space. This means that we can define a nonparametric Bayesian model from a finite-dimensional distribution by simply verifying that (17) holds. From (3), it then immediately follows that the definition (5) uniquely defines a stochastic process and, therefore, SkewGP is a well-defined stochastic process.
is the marginal likelihood of a SkewGP posterior corresponding to the augmented dataset X = [X T , x * T ] T ,ŷ = [y T , 1/2] T . Therefore, we have that , whereγ * ,Γ * are the corresponding matrices of the posterior computed for the augmented dataset.
Proposition 2 This follows by the Fréchet inequality: where A i are events. In fact, note that Φ s+n (γ;Γ ) = P r(u 1 ≥γ 1 , . . . , u s+n ≥γ s+n ) where P r is computed w.r.t. the PDF of a multivariate distribution with zero mean and covarianceΓ .
Theorem 2 The proof follows straightforwardly from that of Corollary 2.

A.1 Additional image classification examples
We have defined a binary sub-problem from the German Traffic Sign data by considering Speed-Limit 30 vs. 50 and from MNIST digit dataset by considering 3 vs. 5. We have compared GP-L vs. SkewGP 2 (both with RBF kernel). Table 2 reports the average accuracy and shows again SkewGP 2 outperforms GP-L (GP-EP achieves lower accuracy than GP-L in both cases).  Our goal is assessing the accuracy but also the quality of the probabilistic predictions. Figure 12, plot (a1), shows, for each instance in the test set (one of the fold) of the MNIST datatset , the value of of the mean predicted probability of class 5 for SkewGP 2 (x-axis) and GP-L (y-axis). Each instance is represented as a blue point. The mean predicted probability ranges in [0.41, 0.53] for GP-L and in [0.25, 0.8] for SkewGP 2 . The red points highlight the instances that were misclassified by GP-L (plot (a2) reports the images of some of the misclassified instances included in the rectangle). Figure (b1) shows the same plot, but the red points are now the instances that were misclassified by SkewGP 2 . By comparing (a) and (b), it is evident that SkewGP 2 provides a higher quality of the probabilistic predictions.
SkewGP 2 also returns a better estimate of its own uncertainty. This is showed in plots (c1) and (c2). For each instance in the test set and for each sample from the posterior, we have computed the predicted class (the class that has probability greater than 0.5). For each test instance, we have then computed the standard deviation of all these predictions and used it to color the scatter plot of the mean predicted probability. In this way, we have a visual representation of first order (mean predicted probability) and second order (standard deviation of the predictions) uncertainty. Plot 12(c1) is relative to GP-L and shows that GP-L confidence is low only for the instances whose mean predicted probability is close to 0.5. This is not reflected in the value of the mean predicted probability for the misclassified instances (compare plot (a1) and (c1) and note that the red spot in (a1) is outside the yellow area in (c1)). Conversely, the second order uncertainty of SkewGP 2 is clearly consistent with plot (b1). Figure 13 shows a similar plot for the "German road-sign" dataset. Plot 13(c1) is relative to GP-L and shows that GP-L confidence is low only for the instances whose mean predicted probability is in [0.3, 0.7]. This is not reflected in the value of the mean predicted probability for the misclassified instances (compare plot (a1) and (c1) and note that some of the red points in (a1) are outside the yellow area in (c1)). Conversely, 13(c2) shows that the second order uncertainty of SkewGP 2 is clearly consistent with plot (b1).