# Discrete Ziggurat: A Time-Memory Trade-Off for Sampling from a Gaussian Distribution over the Integers

## Abstract

Several lattice-based cryptosystems require to sample from a discrete Gaussian distribution over the integers. Existing methods to sample from such a distribution either need large amounts of memory or they are very slow. In this paper we explore a different method that allows for a flexible time-memory trade-off, offering developers freedom in choosing how much space they can spare to store precomputed values. We prove that the generated distribution is close enough to a discrete Gaussian to be used in lattice-based cryptography. Moreover, we report on an implementation of the method and compare its performance to existing methods from the literature. We show that for large standard deviations, the Ziggurat algorithm outperforms all existing methods.

### Keywords

Lattice-based cryptography Gaussian sampling Practicality Implementation## 1 Introduction

The object of study of this paper is the discrete Gaussian probability distribution over the integers. Sampling elements from such a distribution is widely used in lattice-based cryptography [GPV08, LP11, BGV12, GGH12]. It is a critical technical challenge to sample from a discrete Gaussian over the integers accurately and efficiently. Weiden et al. [WHCB13] report that sampling from it takes more than 50 % of the running time of the signing algorithm in their implementation of Lyubashevsky’s signature scheme [Lyu12].

All existing methods to sample from a Gaussian distribution over the integers either need large amounts of memory or they are very slow. For example, Galbraith and Dwarakanath estimate that Peikert’s sampler [Pei10] requires around 12 MB of storage [GD12] for some parameters. Such a large memory requirement might be acceptable on a PC but not on the diversity of devices that demand cryptographic solutions today.

In this paper we explore a different alternative for sampling from a Gaussian distribution over the integers that offers a flexible trade-off between speed and memory. Moreover, for big standard deviations, this method beats commonly used methods. We call the method discrete Ziggurat because it adapts the Ziggurat algorithm [MT00] for the discrete case.

The discrete Ziggurat is specially appealing for cryptographic applications because of its flexibility. The method uses precomputed rectangles of equal ‘size’ to cover the area under the probability density function (PDF). Increasing the number of rectangles increases speed but also increases the memory used. Therefore, it offers an easy-to-tune trade-off between speed and memory. This is a desirable property because developers of cryptographic primitives can easily adjust it to fit the particular characteristics of different devices. On memory constraint devices like smartcards or microcontrollers they could use a low-memory low-speed setting, while on a high performing server they could use a high-memory high-speed configuration.

Originally, the Ziggurat sampler was developed for a continuous distribution. In order to adapt it to the discrete case some care must be taken. In particular the notion of ‘size’ of a rectangle must be redefined from the narrow concept of ‘area’ to the more general “probability to sample points inside the rectangle”. We discuss the implications of this generalization.

It is also challenging to analyze the quality of the discrete Ziggurat because of the subtleties of an actual implementation. In this paper we provide a careful analysis that takes into consideration the loss of precision due to the tailcut, the precision in sampling from the \(y\)-axis and the precision in calculating the PDF. The techniques used and the way they are combined in this analysis might show valuable for the analysis of other samplers. For developers we explain how to achieve a desired accuracy by setting the precision for representing numbers.

We implemented the discrete Ziggurat in C++ using the Number Theory Library (NTL) [Sho]. The implementation can be downloaded at the authors’ homepage^{1}. We compare the efficiency of the discrete Ziggurat with existing methods and analyze the speed-memory trade-off. For example, we used the parameters proposed by Galbraith and Dwarakanath [GD12] for the normal distribution in Lyubashevsky’s signature scheme [Lyu12]. For this illustrative setting, the discrete Ziggurat produces about 1.13 million samples per second, using only 524 KB of memory. In comparison, Peikert’s sampler outputs 281,000 samples per second for a memory usage of 33.55 MB. The Knuth-Yao algorithm is only slightly faster (it produces about 4 % more samples), but increases the memory-consumption by a factor of more than 400.

*Related Work.* We briefly survey existing alternatives to sample from a discrete Gaussian probability distribution over the integers, denoted \(D_{\sigma }\). For parameter \(\sigma >0\), \(D_\sigma \) assigns \(x\in \mathbb {Z}\) a probability proportional to \(\rho _\sigma (x) = \exp (-\frac{1}{2}x^2/\sigma ^2)\). It is important to note that sampling from \(D_\sigma \) is different to sampling from a (continuous) normal distribution [TLLV07]. Another related problem is that of sampling from a Gaussian distribution over a generic lattice, a more complex problem, whose solutions often require sampling from \(D_\sigma \) as a subroutine [GPV08, Pei10, DN12, AGHS12].

For cryptographic applications it is sufficient to sample from the bounded subset \(B:=\mathbb {Z}\cap [-t\sigma ,t\sigma ]\), where the tailcut \(t>0\) is chosen large enough to guarantee a desired precision [GPV08]. One alternative to sample from \(D_\sigma \) is to do rejection sampling on \(B\). Another alternative is to precompute the cumulative distribution function (CDF) for \(x\in B\), sample a uniform element \(y\in [0,1)\) and perform a binary search on the CDF table to output the “inverse CDF” of \(y\) [Pei10]. To the best of our knowledge, no work analyzes the accuracy or efficiency of any of these methods in detail.

Yet another alternative, explored by Galbraith and Dwarakanath [GD12], is the Knuth-Yao algorithm. The algorithm first precomputes a binary tree with leaves labeled by the elements of \(B\). For \(x\in B\), if the probability of sampling \(x\) has a one in the \(i\)-th place of its binary representation, there is a leaf labeled \(x\) at height \(i\) of the tree. Then it samples by walking down the tree using one uniform bit at each step to decide which of the two children to move to. Galbraith and Dwarakanath present a very detailed analysis of the accuracy of the sampler and of the number of random bits it uses. They also propose ways to reduce the memory requirements. However, they do not assess the speed of the sampler.

Ducas and Nguyen propose an enhancement for rejection sampling. They observe that the sampler can compute at a low precision by default and only use high precision computation when a certain threshold is reached [DN12]. To the best of our knowledge, no work evaluates the effect of this enhancement in detail.

*Organization.* In Sect. 2 we explain the Ziggurat algorithm and we describe in detail its discrete variant. In Sect. 3 we analyze the quality of the distribution. Finally, in Sect. 4 we report on experimental results.

## 2 The Discrete Ziggurat Algorithm

The Ziggurat algorithm belongs to the class of rejection sampling algorithms and was introduced by Marsaglia and Tsang for sampling from a continuous Gaussian distribution [MT00]. Here we adapt it for the discrete case. After explaining the setting, we give a short overview over Ziggurat in the continuous case and shortly explain how to control the trade-off. Afterwards, we show how we adapted it to the discrete case and explain how to perform the necessary precomputing. Subsequently, we discuss the implementation-details and finish the section with further improvements.

### 2.1 Setting

We are concerned with sampling from a discrete Gaussian distribution centered at zero with bounded support \(B:=[-t\sigma ,t\sigma ]\cap \mathbb {Z}\) for some parameter \(t>0\). This bounded support is sufficient for the application in lattice-based cryptography as long as \(t\) is chosen large enough. Moreover, we show in Sect. 3.2 how to select parameters such that the sampled distribution is within a certain statistical distance to a (truly) discrete Gaussian distribution. The assumption that the distribution is centered at zero is also fine, as we can add a constant offset to transform samples into a distribution centered around any other integer.

### 2.2 Intuition

Now, how do we sample \(x\) within \(\mathbb {R}^+_0\)? During set-up, we enclose the area of the probability density function (PDF) in an area \(A\) consisting of \(m\) horizontal rectangles with equal area as shown in Fig. 1. How the rectangles are computed is described below. Next, we store the coordinates \((x_i,y_i)\) of the lower right corner of each rectangle \(R_i\), \(1 < i < m-1\). Please note that each rectangle \(R_i\) can be split into a left rectangle \(R_i^l\) that lies completely within the area of the PDF and a right rectangle \(R_i^r\) that is only partially covered by the PDF. For an example, see \(R_3\) in Fig. 1.

Now, to actually sample a value \(x \le t\sigma \) within \(\mathbb {R}^+_0\) we first sample an integer \(1 \le i \le m\) uniformly at random, to select a random rectangle. Next, we sample an \(x\)-coordinate inside \(R_i\), by sampling a uniformly random \(x'\) within \([0,x_i]\). If \(x' \le x_{i-1}\), i.e. if \(x'\) is inside \(R_i^l\), we directly accept and return \(x'\). Otherwise, \(x'\) lies within \(R_i^r\). In this case, we do rejection sampling. Namely, we sample a value \(\gamma \) within \([y_{i+1},y_i]\) uniformly at random. Then, if \(\gamma + y_{i+1} \le \rho _\sigma (x')\), i.e. we hit a point in the area below the PDF, we accept and return \(x'\). Otherwise, we reject and restart the whole process by sampling a new \(i\).

In order to understand why this sampling-algorithm works, think of it as an efficient implementation of rejection-sampling in the area \(A\). More precisely, the implementation of the first step (sampling a point in the enclosing area) is improved. Since all the rectangles have equal size, the probabilities of sampling a point in a given rectangle are equal. Therefore, one can sample the rectangle first and a point in the rectangle afterwards.

The expensive part of the algorithm is computing \(\rho _\sigma (x')\) if \(x'\) does not lie within \(R_i^l\). It becomes even more expensive whenever a value is rejected. For this reason Ziggurat provides a time-memory trade-off, which is controlled by the number of rectangles used, as follows. If we use more rectangles, the ratio between the left and the right rectangle within one rectangle is changed in such a way that the left rectangle becomes comparatively bigger. Hence, we accept an \(x'\) without computing \(\rho _\sigma (x')\) with higher probability. Moreover, using more rectangles, the area \(A\) tighter encloses the area \(C\) below the PDF. Thereby, the area \(A \setminus C\) that leads to a rejection shrinks and with it the overall probability of a rejection. However, for each additional rectangle the coordinates of one additional point have to be stored, increasing the memory requirements.

### 2.3 Adaption to the Discrete Case

In the discrete case, the algorithm works quite similar. The whole pseudocode can be found in Appendix A.1. As before, a sign \(s\), a rectangle with index \(i\) and a potential sample \(x'\) are sampled. If \(x'\) lies in a left rectangle and is non-zero, \(sx'\) is returned immediately. If \(x'\) equals zero, it is returned with probability \(1/2\), like in the continuous case. If not, exactly the same rejection sampling procedure as in the continuous case is used to decide whether \(sx'\) is returned or the whole process is restarted.

In contrast to the continuous case, the notion of ‘size’ defined using the area of a rectangle can not be used in the discrete case. We have seen in the last section that the size of a rectangle has to be proportional to the probability to sample a point in it. In the discrete case, we therefore define the size of a rectangle as the number of integer \(x\)-coordinates in the rectangle times its height. For instance, the rectangle \(R_3\) has size \((1 + \left\lfloor {x_3}\right\rfloor ) \cdot (y_2 - y_3)\).

### 2.4 Implementation

For an implementation, we have to analyze the effect of computing with limited precision. We use a dash over numbers or functions to indicate the use of their corresponding \(n\)-bit fixed-point approximation, e.g. \(\overline{y}\) and \(\overline{\rho }_\sigma \) denote the \(n\)-bit approximation of \(y\in \mathbb {R}\) and the function \(\rho _\sigma \), respectively. Since we can exactly calculate \(\overline{\rho }_\sigma \), we can find a partition such that the rectangles have exactly the same ‘size’ and represent it with the vertical bounds \(\overline{y}_{i}\) (which we store with \(n\) bits fixed point precision) and the rounded horizontal borders \(\left\lfloor {x_i}\right\rfloor \). The last problem is to sample uniformly at random in the infinite sets \([\overline{y}_i,\overline{y}_{i-1}]\). Our solution is to discretize the set: We define \(\overline{h}_i:=\overline{y}_{i-1} - \overline{y}_i\) to be the height of the \(i\)-th rectangle, sample \(y'\mathop {\leftarrow }\limits ^{\$}\{0,1,\dots ,2^{\omega }-1\}\) for a parameter \(\omega \in \mathbb {Z}_0^+\) and transform the samples to \(\overline{y}=\overline{h}_i y'\in [0,2^\omega \overline{h}_i]\). Instead of transforming \(\overline{y}\) into the interval \([\overline{y}_i,\overline{y}_{i-1}]\) we replace the condition \(\overline{y}\le \overline{\rho }_\sigma (x)\) for \(\overline{y}\in [\overline{y}_i,\overline{y}_{i-1}]\) with \(\overline{y}\le 2^\omega (\overline{\rho }_\sigma (x) - \overline{y}_i)\) for \(\overline{y}\in [0,2^\omega \overline{h}_i]\). We show in Sect. 3 how to choose the parameters \(t\), \(\omega \) and \(n\) in order to bound the statistical distance between the distribution defined by our algorithm and \(D_\sigma \) by a given value.

### 2.5 Further Improvement

Since the most time-consuming part of the discrete Ziggurat is the computation of \(\overline{\rho }_\sigma \), we want to avoid it as often as possible. As mentioned above, it is only necessary if \((x,y)\) is contained in a right rectangle \(R_{i}^{r}\). But even in this case, depending on the shape of \(\overline{\rho }_{\sigma }\) inside of \(R_{i}^{r}\), we can avoid the evaluation of \(\overline{\rho }_\sigma \) in nearly half of the cases and more easily reject or accept \(x\) as follows.

## 3 Quality of Our Sampler

In this section, we show how to choose parameters for the algorithm such that it achieves a given quality in the sense of statistical distance to a discrete normal distribution. We begin with a theorem that bounds the statistical distance between the distribution produced by the sampler and a discrete normal distribution. Afterwards, we show as an example how to select parameters such that the statistical distance is smaller than \(2^{-100}\). The methodology can be used to select parameters for any given statistical distance.

### 3.1 Statistical Distance Between Sampled and Gaussian Distribution

No practical algorithm outputs samples exactly distributed according to a Gaussian distribution. Therefore, it is important to understand how much the produced output distribution differs from the normal distribution. This difference is measured by the statistical distance. Recall that \(t\) determines the tailcut and \(\omega \) the precision of the sampled \(y\)-values. As explained before, we use numbers with \(n\)-bit fixed-point precision. Similar to the definition of the support \(B_0^+= [0,t\sigma ] \cap \mathbb {Z}_0^+\), we define \(B^+:= [0,t\sigma ] \cap \mathbb {Z}^+\). The next theorem gives a lower bound on the quality of our sampler depending on the used parameters.

**Theorem 1**

Because of the restricted space, we omit the proof of the result here. It can be found in the full version of this paper^{2}. The main idea of the proof is to introduce intermediary distributions. The first intermediary distribution differs from a Gaussian distribution by the tailcut. The second intermediary distribution takes the limited precision of the stored numbers and the sampled \(y\)-values into consideration. After bounding the statistical distances between the consecutive distributions, we apply the triangle inequality to show the main result.

### 3.2 Parameter Selection

## 4 Experiments and Results

In this section we discuss the performance of our implementation of the discrete Ziggurat algorithm. We first describe the experiments we performed to test the efficiency, then present their results and analyze the gathered data. Furthermore, we compare our sampler to implementations of existing samplers for discrete Gaussians.

### 4.1 Environment and Setup

The experiments were carried out on a Sun XFire 4400 server with 16 Quad-Core AMD Opteron 8356 CPUs running at 2.3 GHz (we only used one CPU), having in total 64GB of RAM and running a 64 bit Debian 7.1. All implementations use the Number Theory Library (NTL, cf. [Sho]) with precision \(n=106\) bits in consistency to our choice of parameters in Sect. 3.2 to assure a statistical distance for Ziggurat of at most \(2^{-100}\). Furthermore, we used the tailcut \(t=13\) and the discrete Gaussian parameters \(\sigma \in \{ 10, 32, 1000, 1.6\cdot 10^{5} \}\). The value \(\sigma =32\) maintains the worst-to-average-case reduction [Reg05] in several schemes for a certain parameter set, and the rather large value \(\sigma = 1.6\cdot 10^{5}\) is chosen according to Galbraith and Dwarakanath [GD12]. The other two values \(\sigma =10,1000\) were chosen arbitrarily inbetween and at the lower end to allow a better comparison.

We queried each algorithm iteratively 1 million times to output a single sample per call. These experiments were applied to the discrete Ziggurat with the optimization using the straight line \(s\) (ZigguratO), discrete Ziggurat without optimization (Ziggurat), inverse CDF (invCDF), rejection sampling with precomputed lookup-table (rejSam), and Knuth-Yao (KY). Furthermore we tested both Ziggurat algorithms with a precomputed lookup-table for the support \(B_{0}^{+}\) (ZigguratOP and ZigguratP, respectively).

For each algorithm we measured the running time using the (Linux-internal) function clock_gettime with clock CLOCK_PROCESS_CPUTIME_ID. In order to have non-distorted results we excluded all pre- and post-computations (e.g. setting up lookup-tables) from the measurements. Regarding the memory, we did not perform per-runtime analyses, but computed the amount of memory by adding up the number of fixed variables in regard to their types in NTL. For our choice of parameters, in Ziggurat(O) the values on the \(x\)-axis need 8 bytes and on the \(y\)-axis 24 bytes of memory. With \(m\) rectangles the total amount of memory is thus \(32 (m+2)\) bytes (including \(\sigma , t, \omega , m\)). For both invCDF and rejSam we need to store a lookup-table of \(t\sigma \) values à 16 bytes, resulting in 2080 bytes for \(\sigma =10\). The same amount of memory is used by Ziggurat(O) with \(m=63\) rectangles. The size of Knuth-Yao is approximated by \((\#\text {intermediates}+\#\text {leaves})/2\) bits, where \(\#\text {intermediates} = n\cdot 2^{\left\lceil {\log \log (n\cdot t\sigma )}\right\rceil }\) and \(\#\text {leaves} = n\cdot 2^{\left\lceil {\log \log (t\sigma )}\right\rceil }\) for precision \(n=106\) bits as above.

### 4.2 Results

Figure 3 shows results for inverse CDF, rejection sampling, Knuth-Yao, and discrete Ziggurat with and without optimizations for different numbers of rectangles. It shows four different graphs for different values of \(\sigma \). For small values of \(\sigma \), the inverse CDF method outperforms both discrete Ziggurat and rejection sampling for the same fixed amount of memory. For example, our implementation invCDF samples about 1.37 million samples per second for \(\sigma =32\). On the other hand, rejection sampling is quite slow due to a large rejection area. Even with a precomputed lookup-table, rejSam only achieves about 327,000 samples per second, which is a factor \(4.2\) slower than invCDF. The naïve approach without lookup-table solely achieves 2,500 samples per second, being a factor \(558\) slower than invCDF. For the same amount of memory, ZigguratO achieves an overall number of about 753,000 samples per second, while Ziggurat outputs 718,000 samples per second. Compared to the other two methods, Ziggurat is 1.91 times slower than invCDF and a factor 2.19 faster than rejSam. Our implementation of Knuth-Yao outperforms all the other methods by at least a factor of 3.53, outputting 4.85 million samples per second. This comes at the cost of nearly doubled memory usage.

In the extreme case \(\sigma =1.6\cdot 10^{5}\), the fastest instantiation of Ziggurat outputs 1.13 million samples per second with a memory usage of 524 KB. Inverse CDF creates 281,000 samples per second while using 33.55 MB, thus being about a factor 4 slower than Ziggurat. For rejSam the situation is even worse: Using the same amount of memory as invCDF, it only outputs 185,000 samples per second – a factor 6.1 slower than Ziggurat. The Knuth-Yao algorithm still performs better than Ziggurat, but only by 4.26 %. On the other hand, KY needs more than 424 times the memory storage of Ziggurat. Concluding we state that for larger values of \(\sigma \) the Ziggurat algorithm beats both inverse CDF and rejection sampling. Compared to Knuth-Yao, Ziggurat achieves almost the same speed but reduces the memory consumption by a factor of more than 400.

Figure 3 shows that we can beat invCDF in terms of speed and compete with Knuth-Yao. The reason for this is the simplicity of the algorithm. If many rectangles are stored, the rejection-probability gets very small. Likewise, the probability to sample an \(x\) in a right rectangle \(R_i^r\) gets very small. Therefore, the algorithm only samples a rectangle and afterwards samples a value within this rectangle, which can be done very fast.

The trade-off provided by the Ziggurat-algorithms is indeed a property the other approaches do not share. InvCDF assigns every possible value to an interval on the \(y\)-axis. Consequently, one has to store at least the borders if the intervals. Decreasing the precision of the borders will decrease the memory consumption, but as well decrease the quality of the sampler. Increasing the precision or storing intermediate values, on the other hand, will not decrease the running time. The same happens to rejection sampling if the precision of the precomputed values is changed. Knuth-Yao stores for every element in the support the probability to sample this element. Decreasing the precision of the stored probabilities would (like for invCDF) decrease the quality of the sampler. While there might be efficient ways to store those values, there is a minimal amount of space required to store this information. Knuth-Yao as well as invCDF and rejection sampling therefore only provide a trade-off between quality and speed/memory consumption.

In Fig. 4 we draw the time-memory trade-off for the Ziggurat algorithm for different values of \(\sigma \). One can see that the performance of the Ziggurat algorithm decreases for larger \(\sigma \). What is interesting in the graph is that the Ziggurat algorithm for \(\sigma =10\) is slower for a large amount of rectangles than for \(\sigma =32\). This is puzzling as we cannot directly explain the described behaviour. We want to state that during our experiments we saw quite large fluctuations for several runs of the algorithm. Maybe this could explain the better performance for \(\sigma =32\) in comparison to \(\sigma =10\).

We also compared ZigguratO and Ziggurat in regard to speed.^{3} The improvement rate increases up to 30 % for a total memory of 320 bytes, then decreases to around 6 % for 2080 bytes, until finally for 130KB and bigger there is no improvement. Overall, the described behaviour is not surprising since for increased memory the number of rectangles gets larger, so that the rejection area is very small. This leads to nearly no evaluations in the right sub-rectangles \(R_{i}^{r}\) and therefore to no computation of the straight line \(s\) (or even \(\overline{\rho }_{\sigma }\)).

Additionally, we compared ZigguratO to ZigguratOP, which operates with at least 2176 bytes of memory. ZigguratOP is slower until about 2.5 KB of memory, but then it beats ZigguratO with a speedup of up to 40 %, until for memory larger than 262 KB there seems to be no speedup at all. This behaviour is reasonable since the lookup-table requires more storage, but simultaneously affects the speed due to replacing \(\overline{\rho }_{\sigma }\) by table-lookups.

## Footnotes

- 1.
- 2.
The full version is available at http://eprint.iacr.org/2013/510.pdf.

- 3.
For additional Figures see Appendix A.2.

### References

- [AGHS12]Agrawal, S., Gentry, C., Halevi, S., Sahai, A:. Discrete Gaussian leftover hash lemma over infinite domains. Cryptology ePrint Archive, Report 2012/714 (2012). http://eprint.iacr.org/
- [BGV12]Brakerski, Z., Gentry, C., Vaikuntanathan, V.: (Leveled) fully homomorphic encryption without bootstrapping. In: Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, ITCS ’12, pp. 309–325. ACM, New York (2012)Google Scholar
- [DN12]Ducas, L., Nguyen, P.Q.: Faster Gaussian lattice sampling using lazy floating-point arithmetic. In: Wang, X., Sako, K. (eds.) ASIACRYPT 2012. LNCS, vol. 7658, pp. 415–432. Springer, Heidelberg (2012) CrossRefGoogle Scholar
- [GD12]Galbraith, S.D., Dwarakanath, N.C.: Efficient sampling from discrete Gaussians for lattice-based cryptography on a constrained device. Preprint (2012)Google Scholar
- [GGH12]Garg, S., Gentry, C., Halevi, S.: Candidate multilinear maps from ideal lattices. Cryptology ePrint Archive, Report 2012/610 (2012). http://eprint.iacr.org/
- [GPV08]Gentry, C., Peikert, C., Vaikuntanathan, C.: Trapdoors for hard lattices and new cryptographic constructions. In: Ladner, R.E., Dwork, C. (ed.) 40th ACM STOC Annual ACM Symposium on Theory of Computing, Victoria, British Columbia, Canada, 17–20 May 2008, pp. 197–206. ACM Press (2008)Google Scholar
- [LP11]Lindner, R., Peikert, Ch.: Better key sizes (and attacks) for LWE-based encryption. In: Kiayias, A. (ed.) CT-RSA 2011. LNCS, vol. 6558, pp. 319–339. Springer, Heidelberg (2011) Google Scholar
- [Lyu12]Lyubashevsky, V.: Lattice signatures without trapdoors. In: Pointcheval, D., Johansson, T. (eds.) EUROCRYPT 2012. LNCS, vol. 7237, pp. 738–755. Springer, Heidelberg (2012) CrossRefGoogle Scholar
- [MT00]Marsaglia, G., Tsang, W.W.: The Ziggurat method for generating random variables. J. Stat. Softw.
**5**(8), 1–7, 10 (2000)Google Scholar - [Pei10]Peikert, Ch.: An efficient and parallel Gaussian sampler for lattices. In: Rabin, T. (ed.) CRYPTO 2010. LNCS, vol. 6223, pp. 80–97. Springer, Heidelberg (2010) CrossRefGoogle Scholar
- [Reg05]Regev, O.: On lattices, learning with errors, random linear codes, and cryptography. In: Proceedings of the Thirty-Seventh Annual ACM Symposium on Theory of Computing, STOC ’05, pp. 84–93. ACM, New York (2005)Google Scholar
- [Sho]Shoup, V.: Number Theory Library (NTL) for C++. http://www.shoup.net/ntl/
- [TLLV07]Thomas, D.B., Luk, W., Leong, P.H.W.: Gaussian random number generators. ACM Comput. Surv.
**39**(4), 11:1–11:38 (2007)CrossRefGoogle Scholar - [WHCB13]Weiden, P., Hülsing, A., Cabarcas, D., Buchmann, J.: Instantiating treeless signature schemes. Cryptology ePrint Archive, Report 2013/065 (2013). http://eprint.iacr.org/