1 Introduction

“Long memory’, or ’long range dependence’, or ’strong dependence’, in time series is a topic that has been quite extensively studied in recent years. After a number of probabilistic contributions and empirical studies, serious treatment of issues of statistical inference could be said to have begun in the mid-1980s, with activity then increasing through the 1990s and this century. Predominately, this literature has focussed on observations that are regularly spaced over time, and the bulk of the theoretical development has been in terms of asymptotic statistical theory, with the number of observations, n, regarded as diverging, finite-sample theory proving mathematically intractable, even under the precise distributional assumptions that are typically not required in a large-sample treatment. Parametric, semiparametric and nonparametric models have all featured. The major characteristic feature of a long memory covariance stationary time series process is that its autocovariance function decays so slowly with increase in lag length as not to be summable, or, nearly equivalently, its spectral density diverges, typically at zero frequenccy; while some nonstationary processes, such as ones with a unit root can, a fortiori, be regarded as having even longer memory. By contrast, ’short memory’ time series typically have autocovariances that are summable and spectral density that is more or less smooth (for example a stationary autoregressive moving average (ARMA) has exponentuially decaying autocovariance and analytic spectral density); though for some relevant purposes, a short memory process is sometimes defined as merely having spectral density that is positive and finite at zero frequency.

Spatial data have long attracted the attention of statisticians, and the configurations of some such data, especially some arising in such fields as meteorology, cosmology and agriculture, can be viewed as generalizations of the typical regularly spaced time series one mentioned above. In particular, they constitute observations in \(M\ge 2\) dimensions, (where \(M=1\) in the tme series case), such that there are \(n_{m}\)observations across dimension m,  for \(m=1,\ldots ,M,\) so \(n=\Pi _{m=1}^{M}n_{m}\). We call this a ’(rectangular) lattice’, and data recorded on it belong to the class of ’random fields’. There is equal spacing across each dimension, but the spacing can vary across dimensions. In Fig. 1 for (horizontal) dimension \(1,n_{1}=15,\) while for (vertical) dimension 2, \(n_{2}=7.\) When \(n_{1}=n_{2}\) we can speak of a ’square lattice’.

Fig. 1
figure 1

Spatial lattice

Figure 1 brings to mind regularly spaced agricultural plots on a field. In fact, ‘spatial long memory’ goes back at least to the agriculturally motivated paper of Smith (1938), which is also a very early reference relative to the literature on time series long memory. Consider \(n=n_{1}n_{2}\) agricultural plots on a field, as in Fig. 1. Denote the yield (of rice, say) at location \(i=\left( i_{1},i_{2}\right) \) by \(x_{i_{1}i_{2}}\) and the sample mean by \(\overline{x}=n^{-1}\Sigma _{i_{1}=1}^{n_{1}}\Sigma _{i_{2}=1}^{n_{2}}x_{i_{1}i_{2}}\). Smith (1938) assumed in effect that

$$\begin{aligned} \mathrm{Var}\left( \overline{x}\right) \sim Cn^{2d-1},0\le d<1/2,C>0, \end{aligned}$$
(1)

and estimated the unknown d and C from the logged relation

$$\begin{aligned} \log \left( \mathrm{Var}\left( \overline{x}\right) \right) \sim \log C+ \left( 2d-1\right) \log n, \end{aligned}$$

meaning that the left side is estimated from data over several fields. Now, in the special square lattice case, \(n_{1}=n_{2}=n^{1/2},\) the model (1) is consistent with the underlying isotropic model

$$\begin{aligned} \mathrm{Cov}\left( x_{i_{1}i_{2}},x_{j_{1}j_{2}}\right) \sim c\sqrt{\left( i_{1}-j_{1}\right) ^{2}+\left( i_{2}-j_{2}\right) ^{2}}^{2d-1}=c\left\| i-j\right\| ^{2d-1}, \end{aligned}$$

for \(i=\left( i_{1},i_{2}\right) ,\) \(j=\left( j_{1},j_{2}\right) ,\) which is equivalent to the usual autocovariance function of a stationary long memory time series with differencing or memory parameter d, e.g., a fractional autoregressive integrated moving average (FARIMA). To verify (1), note that

$$\begin{aligned} \mathrm{Var}\left( \overline{x}\right)= & {} n^{-2}\sum \limits _{i_{1},i_{2}}\sum \limits _{j_{1},j_{2}}\mathrm{Cov}\left( x_{i_{1}i_{2}},x_{j_{1}j_{2}}\right) \\\sim & {} cn^{-2}\sum \limits _{i_{1},i_{2}} \sum \limits _{j_{1},j_{2}}\sqrt{\left( \left( i_{1}-j_{1}\right) ^{2}+\left( i_{2}-j_{2}\right) ^{2}\right) ^{2d-1}} \\= & {} cn^{-2}\sum \limits _{h,k} \left( n^{1/2}-\left| k\right| \right) \left( n^{1/2}-\left| h\right| \right) \sqrt{h^{2}+k^{2}} ^{2d-1} \\\sim & {} Cn^{2d-1}. \end{aligned}$$

It is interesting that Smith (1938) thought of a power law decay: this might seem natural to one coming to the subject unschooled in time series modelng which, after World War II, stressed exponential decay, as in ARMA modeling.

Since then, many papers on ’spatial long memory’ have appeared, but the topic has not been developed as systematically or comprehensively as ’long memory time series’. Some distinctive issues arising in the ’spatial’ case, all of which have been studied far more under short memory than long memory, are:

  1. 1.

    Is there isotropy or not? If not, we might model each dimension separately, or with some interaction, and possibly with a different memory parameter for each dimension.

  2. 2.

    Is sampling regularly or irregularly spaced? Whereas in time series regular spacing has been predominately studied, irregular spacing is perhaps more likely to be found with spatial data.

  3. 3.

    Should modeling be unilateral or multilateral? For time series unilateral modeling, reflecting one-sided transition from past to future, is usually natural, but this is not the case with spatial data, where, for example, the dimensions might be latitude and longitude.

  4. 4.

    The edge effect. In estimating lagged quantities, there is loss of information at the boundary of the observation region, which has negligible effect when \(M=1\), but increasing, and damaging, effect as M increases unless corrected for.

In this, quite personal, view of the subject, we discuss the following topics in the remaining three sections of the paper:

  1. 1.

    Inference on location and regression with long memory errors.

  2. 2.

    Inference on second-order properties of long memory stationary processes.

  3. 3.

    Miscellaneous topics: nonstationary processes, irregular spacing, adaptive estimation, nonparametric regression.

We do not consider ’spatial autoregressive’-type (’SAR’) models (which depend on a user-chosen weight matrix of geographic or economic inverse distances); these do not fit into our framework but typically possess a kind of short memory.

2 Inference on location and regression with long memory errors

The sample mean \(\overline{x}\) is a basic statistic, whose asymptotic properties are well known under time series short and long memory. In the time series case, suppose

$$\begin{aligned} x_{i}=\mu +\sum \limits _{j=0}^{\infty }\alpha _{j}\varepsilon _{i-j}, \sum \limits _{j=0}^{\infty }\alpha _{j}^{2}<\infty , \end{aligned}$$
(2)

where \(\left\{ \varepsilon _{j}\right\} \) is a sequence of independent and identically distributed (iid) random variables with zero mean and unit variance, or even homoscedastic (but not necessarily identically distributed) martingale differences. The model (2) implies \(x_{t}\) is covariance stationary, and it can have ‘short memory’ or ‘long memory’ or ‘negative memory’.

First suppose \(x_{i}\) has short memory, i.e., \(x_{i}\) has spectral density

$$\begin{aligned} f\left( \lambda \right) =\frac{1}{2\pi }\sum \limits _{j=-\infty }^{\infty }\mathrm{Cov}\left( x_{i},x_{i+j}\right) \cos \left( j\lambda \right) ,\quad -\pi \le \lambda \le \pi , \end{aligned}$$

that satisfies

$$\begin{aligned} 0<f\left( 0\right) <\infty . \end{aligned}$$

with also \(f\left( \lambda \right) \) continuous at \(\lambda =0\). Then, \( \overline{x}=n^{-1}\sum \nolimits _{i=1}^{n}\) \(x_{i}\) is an asymptotically normal and efficient estimate of \(\mu \),

$$\begin{aligned} \sqrt{n}\left( \overline{x}-\mu \right) \rightarrow _{d}\mathcal {N}\left( 0,2\pi f\left( 0\right) \right) , \quad \text {as }n\rightarrow \infty . \end{aligned}$$

Long memory can be described by (2) with, for example,

$$\begin{aligned} \alpha _{j}\sim \ell (j)j^{d-1}, \quad 0<d<1/2, \end{aligned}$$
(3)

for a slowly varying function \(\ell (j)\) so \(f\left( 0\right) =\infty \). We can also consider the asymptotic approximation in (3) for \(-1/2<d<0,\) assuming also \(\sum \nolimits _{j=0}^{\infty }\alpha _{j}=0,\) whence \(f\left( 0\right) =0\), and there is ’negative memory’. In both cases, there is interest in allowing for a nonconstant \(\ell (j)\) (e.g., \(\ell (j)=\log j\) ), and this has been done in limit theory for basic statistics, and also in (semiparametric) estimation of d. But for simpliciy, we focus on constant \(\ell (j),\) which covers FARIMA and fractional Brownian motion models. In this setting, \(\overline{x}\) is no longer asymptotically efficient (Adenstedt 1974) but it can still be asymptotically normal, albeit with a different rate of convergence. In particular, under suitable conditions on \(\left\{ \varepsilon _{j}\right\} \),

$$\begin{aligned} n^{1/2-d}\left( \overline{x}-\mu \right) \rightarrow _{d}\mathcal {N}\left( 0,\sigma ^{2}\left( d\right) \right) \quad \text {as }n\rightarrow \infty , 0<\left| d\right| <1/2, \end{aligned}$$

for a function \(\sigma ^{2}\left( d\right) \) of known form.

More generally consider the centered and self-normalized statistic

$$\begin{aligned} u=\left( \overline{x}-\mu \right) /\left\{ E\left( n^{-1}\sum \limits _{i=1}^{n}\left( x_{i}-\mu \right) \right) ^{2}\right\} ^{1/2}. \end{aligned}$$

Ibragimov and Linnik (1971) showed that

$$\begin{aligned} u\rightarrow _{d}\mathcal {N}\left( 0,1\right) ,\text { as }n\rightarrow \infty , \end{aligned}$$

if only the \(\varepsilon _{j}\) are iid and

$$\begin{aligned} \sum \limits _{j=0}^{\infty }\alpha _{j}^{2}<\infty ,E\left( \sum \limits _{i=1}^{n}\left( x_{i}-\mu \right) \right) ^{2}\rightarrow \infty ,\text { as }n\rightarrow \infty . \end{aligned}$$

Note that under negative memory

$$\begin{aligned} E\left( \sum \limits _{i=1}^{n}\left( x_{i}-\mu \right) \right) ^{2} \end{aligned}$$

diverges more and more slowly as \(d\downarrow -1/2\), indeed in the moving average (MA) model with an MA unit root, \(y_{t}=\mu +\varepsilon _{t}-\varepsilon _{t-1},\) we have \(\sum \nolimits _{i=1}^{n}\left( x_{i}-\mu \right) =\varepsilon _{n}-\varepsilon _{0},\) and it does not diverge at all.

For spatial data, consider an \(M-\)dimensional lattice, as described in the previous section (though results can be extended to a more general, irregular region). Let i be the multiple index \((i_{1},\ldots ,i_{M})\) with \( i_{j}\in \mathbb {Z}=\{0,\pm 1,\ldots \}\), \(j=1,\ldots ,M.\) Consider a covariance stationary process \(x_{t}\) observed on the lattice \(\mathbb {N} {=} \{i: \) \(1\le i_{m}\le n_{m}\), \(m=1,\ldots ,M\}\),

$$\begin{aligned} x_{i}=\sum \limits _{j\in \mathbb {Z}^{M}}\alpha _{i-j}\varepsilon _{j}, \sum \limits _{j\in \mathbb {Z}^{M}}\alpha _{j}^{2}<\infty , \end{aligned}$$

for \(\varepsilon _{j}\) iid (say) with zero mean and unit variance. For example under isotropy, with \(M=2,\)

$$\begin{aligned} \alpha _{j}\sim \left\| j\right\| ^{-\delta }=\left( j_{1}^{2}+j_{2}^{2}\right) ^{-\delta /2}, \end{aligned}$$

with memory parameter \(\delta >1\). Note that for time series, long memory has been ’explained’ as arising from cross-sectional aggregation of AR\(\left( 1\right) \) processes, see, Robinson (1978), Granger (1980), an interpretation extended to spatial processes by, e.g., Lavancier (2011).

For \(n=\Pi _{m=1}^{M}n_{m}\) and the multiple partial sum process

$$\begin{aligned} S=\sum \limits _{i\in \mathbb {N}}x_{i}, \end{aligned}$$

Specializing their results to the case \(M=2,\) Lahiri and Robinson (2016) obtained

$$\begin{aligned} n^{\delta -3}S\rightarrow _{d}\mathcal {N}\left( 0,\sigma ^{2}\left( \delta \right) \right) , \end{aligned}$$

with

$$\begin{aligned} \delta\in & {} \left( 1,2\right) \text { (long memory),} \\ \delta\in & {} \left( 2,2.5\right) ,\sum \limits _{j\in \mathbb {Z}^{2}}\alpha _{j}=0 \text { (negative memory),} \end{aligned}$$

and also

$$\begin{aligned} n^{-1/2}S\rightarrow _{d}\mathcal {N}\left( 0,\sigma ^{2}\left( \delta \right) \right) \end{aligned}$$

with

$$\begin{aligned} \delta >2.5,\sum \limits _{j\in \mathbb {Z}^{2}}\alpha _{j}=0. \end{aligned}$$

Such results can be extended to regression models, especially ones with deterministic regressors. One interesting case, for \(M=2,\) and with \(y_{i}\) a response variable that is obserrvable on the lattice is, with \(\beta _{0},\beta _{1},\beta _{2},\) \(\theta _{1},\theta _{2}\) all unknown,

$$\begin{aligned} y_{i}=\beta _{0}+\beta _{1}i_{1}^{\theta _{1}}+\beta _{2}i_{2}^{\theta _{2}}+x_{i}, \end{aligned}$$

with \(x_{i}\) now an unobservable error process. For a more general regression model, and with general M, Robinson (2012) considered nonlinear least squares estimation of the \(\beta ^{\prime }\)s and \(\theta ^{\prime }s,\) for short memory \(x_{i},\) where \(\theta _{1}>-1/2,\) \(\theta _{2}>-1/2\) is required (with negative values of the \(\theta _{i}\) conferring a decaying trend that could be relevant for some spatial data). In this case, the estimates are asymptotically efficient (in the Gauss–Markov sense), but the \( \beta ^{\prime }\)s are estimated slightly less well than the \(\theta ^{\prime }\)s, with rates dependng on the \(\theta ^{\prime }\)s (and slightly less well than if the \(\theta ^{\prime }\)s were known) . With long memory \( x_{i}\), where regression powers will ’interact’ with memory parameters, the rates will be slower, and if memory is strong enough relative to the magnitude of the \(\theta ^{\prime }\)s we may not be able to estimate some \( \beta ^{\prime }\)s and \(\theta ^{\prime }s\), due to domination of the regression component by the error \(x_{i}\).

3 Inference on second-order properties of long memory stationary processes

Now consider estimating either a full parametric model for the spectrum/autocovariance function of \(x_{i}\), or else a ’semiparametric’ one specified for only low frequencies/long lags. In the parametric case, for short memory time series, early work on asymptotics covered least squares and Yule–Walker estimates of stationary autoregressive (AR) processes. Continuous and discrete frequency Whittle, and Gaussian pseudo-maximum likelihood, estimates of ARMA and other short memory time series were well covered by Hannan (1973). He relaxed the commonly imposed iid assumption on innovations, allowing them, and centred squared innovations, to be stationary martingale differences (expressing the natural ordering of time series data), with only finite second moments required (for estimation of dependence parameters); in fact stationarity can be relaxed, the squared innovations needing only to be uniformly integrable). His central limit theorem proof was essentially found to work under long memory with differencing parameter \(d\in \left( 0,1/4\right) \) by Yajima, (1985). Using a different method of proof, Fox and Taqqu (1986) established a central limit theorem valid over the whole stationary long memory interval \(d\in \left( 0,1/2\right) \).

For spatial lattice data, a vital early (short memory) reference is Whittle (1954). Noting that multilateral MA/AR representations are more natural for spatial data than the unilateral or one-sided ones normal in time series, he pointed out identifiability problems with multilateral representations, and extended the (unilateral) Wold representation for purely nondeterministic time series (cf. (1)) to spatial processes, introducing ’half-plane’ representations. But a given multilateral ARMA does not necessarily have a neat half-plane representation, where AR or MA orders may be infinite; so this property may have limited practical use. Sometimes, a ’quarter plane’ representation is possible.

Whittle (1954) and others (e.g., Martin 1979,Tjostheim 1978,Kashyap and Lapsa 1984; Huang and Anh 1992; Ma 2003) considered estimation of various short memory parametric spatial models, some extending ARMA in an isotropic or separable way, e.g., for \(M=2,\)

$$\begin{aligned} \left( 1-\alpha _{1}L_{1}\right) \left( 1-\alpha _{2}L_{2}\right) x_{t}=\varepsilon _{i}, \end{aligned}$$

where \(L_{1}x_{t}=x_{i_{1}-1,i_{2}},\) \(L_{2}x_{t}=x_{i_{1}, i_{2}-1}\) and the \(\varepsilon _{i}\) are iid, a kind of first-order, unilateral AR in 2 dimensions. Other specifications, such as the Matern model, were considered by Stein (1999).

For asymptotic theory of estimates, one basic question is whether to use increasing domain (as usually in time series) or fixed-domain (infill) asymptotics (or some hybrid of the two, where the domain increases slowly while the sampling interval decays slowly). Since spatial observations can often be thought of as confined to a bounded region, infill asymptotics, as studied by Stein (1999), might seem natural. However, infill asymptotics can lead to results that are not easy to use in practice and even to inconsistency, whereas increasing domain asymptotics often provides a central limit theorem that is convenient for practical use.

Another important issue is the ’edge effect’. Consider estimating the autocovariance \(\gamma _{j}\) \(=\mathrm{Cov}\left( x_{i},x_{i+j}\right) \) of a (zero mean) covariance stationary process \(x_{i}\) on an \(M-\)dimensional lattice, by

$$\begin{aligned} \widehat{\gamma }_{j}=\frac{1}{n}\sum \limits _{i,i+j\in \mathbb {N} }x_{i}x_{i+j}. \end{aligned}$$

This is of great interest because various estimates of parameters, in particular ones based on a Gaussian pseudo-likelihood or approximation thereto (and indeed nonparametric spectral estimates) are essentially functions of the \(\widehat{\gamma }_{j}\). For \(M=1\) , as in time series, the summation includes \(n-j\) points and so (for fixed j), \(\widehat{\gamma }_{j}\) has bias \(O\left( n^{-1}\right) ,\) which does not prevent a central limit theorem for \(n^{1/2}\left( \widehat{\gamma }_{j}-\gamma _{j}\right) \) from holding. But for \(M=2,\) and \(\mathbb {N}=\left[ 1,n^{1/2} \right] \times \left[ 1,n^{1/2}\right] \) the bias is of exact order \( n^{-1/2} \) so the limit distribution of \(n^{1/2}\left( \widehat{\gamma } _{j}-\gamma _{j}\right) \) has nonzero mean. And for \(M=3\) and \(\mathbb {N}= \left[ 1,n^{1/3}\right] \times \left[ 1,n^{1/3}\right] \times \left[ 1,n^{1/3}\right] \) the bias is of exact order \(n^{-1/3}\); so, no central limit theorem results, likewise for \(M>3\). Thus, the obvious extensions of time series estimates to such spatial lattice data do not work.

The following solutions, have been proposed, all in a short memory context.

  1. 1.

    (Guyon (1982)) Use unbiased estimates

    $$\begin{aligned} \widetilde{\gamma }_{j}=n\widehat{\gamma }_{j}/\#\left\{ i:i,i+j\in \mathbb {N }\right\} \end{aligned}$$

    of \(\gamma _{j}\). However the \(\widetilde{\gamma }_{j}\) become unstable for large \(\left\| j\right\| \) and a desirable non-negative definiteness property of the \(\widehat{\gamma }_{j}\) is lost.

  2. 2.

    (Dahlhaus and Kuensch (1987)) A data taper, of the type extending that used to correct for bias in a variety of time series settings, is applied to \(x_{i}\).

  3. 3.

    (Robinson and Vidal Sanz (2006); Robinson (2007)) Trimming is employed: omitting the \(\widetilde{\gamma }_{j}\) defined above for large \(\left\| j\right\| \), the first reference considering a parametric setting, the second a nonparametric one.

Because of the greater importance of long lags, i.e., large \(\left\| j\right\| ,\) in long memory processes relative to short memory ones, the edge effect seems an even bigger issue under long memory.

Nevertheless, a number of parametric long memory stationary linear spatial models, both isotropic and separable, have been considered in the literature, see, e.g., Sethuram and Basawa (1995), Boissy et al. (2005), Ghodsi and Shitan (2009), Beran et al. (2009). A simple illustration of a separable long memory spatial parametric model is

$$\begin{aligned} \left( 1-L_{1}\right) ^{d_{1}}\left( 1-L_{2}\right) ^{d_{2}}x_{i}=\varepsilon _{i},\quad 0<d_{1},d_{2}<1/4. \end{aligned}$$

For cyclic/seasonal long memory models, with a spectral pole at one or more non-zero frequencies, see, e.g., Cisse et al. (2016).

For time series ’semiparametric’ estimates of the memory parameter, only local-to-zero assumptions are made about the spectrum, and an increasing number but vanishing fraction of Fourier frequencies near the origin, depending on a user-chosen bandwidth, are used. These have also been extended to spatial lattice data. Log periodogram spatial estimates have been considered by Ghodsi and Shitan (2016), the latter extending asymptotic theory of Robinson (1995a) (though Ghodsi and Shitan (2016) assume a parametric model). Local Whittle or Gaussian semiparametric estimates have been developed by Guo et al. (2009), Durastanti et al. (2014), extending asymptotc theory of Robinson (1995b).

There is considerable scope for for further study of issues of model choice (especially due to the danger of ’curse of dimensionality’ in lattice models) and bandwidth choice. Basic asymptotic theory for estimation of second-order features of long memory lattice processes has been developed by Lavancier (2006, 2007, 2008), and Lavancier and Philippe (2011).

Sometimes, though observations are made only at discrete points in \(\mathbb {Z }^{M}\), it is natural to think in terms of an underlying continuous model defined on \(\mathcal {R}^{M},\) where \(\mathcal {R}\) denotes the real line. This topic was long ago extensively discussed in the time series case \(M=1\), especially for short memory models. The most popular covariance stationary model here has been expressed as a rational spectral density, of a continuous process x(t),  \(t\in \mathcal {R}\),

$$\begin{aligned} g\left( \lambda \right) =\frac{1}{2\pi }\frac{\left| b\left( i\lambda \right) \right| ^{2}}{\left| a\left( i\lambda \right) \right| ^{2}},\lambda \in \mathcal {R}, \end{aligned}$$
(4)

where a and b are polynomials of degrees r and s, respectively such that \(r>s\) (required for integrability of g,  equivalently for x(t) to have finite variance), and such that the zeroes of a have negative real parts. When \(s=0\), we can think of x(t) as generated by a stochastic differential equation (of order r) driven by ’continuous white noise’, that is an unobservable process \(\zeta \left( t\right) ,\) \(t\in \mathcal {R}\), having (non-integrable) spectral density

$$\begin{aligned} h\left( \lambda \right) =\frac{1}{2\pi },\lambda \in \mathcal {R}, \end{aligned}$$
(5)

the b factor in (4) providing further structure when \(r>1\), \(s>0\). Suppose that \(x_{t}=x(t)\) is observable at \(t\in \mathbb {Z}\) only. Then (see, e.g., Bartlett 1946, Walker 1960, Phillips 1959; Robinson 1980a) \(x_{t}\) has an ARMA(\(r,r-1\)) representation. This can be established in more than one way. Note that, if \(f\left( \lambda \right) \) denotes the spectral density of \(x_{t}\),

$$\begin{aligned} f\left( \lambda \right) =\underset{j=-\infty }{\overset{\infty }{\Sigma }} g\left( \lambda +2\pi \right) , \end{aligned}$$
(6)

and Robinson (1980a) thence used (6) with contour integration/theorem of residues to derive the ARMA(\(r,r-1\)) property of \(x_{t},\) also using it to derive the ARMA(rr) property of the averaged process \(X_{t}=\mathop {\smallint }\nolimits _{t-1}^{t}\) \(x(u)\mathrm{d}u,\) \(t\in \mathbb {Z}\) (as well as analogous properties when the underlying process is skip-sampled, being defined discretely at narrow intervals 1 / k,  for an integer \(k\ge 2,\) see also Telser (1967)). Generally, there are efficiency gains if the underlying parameters in a and b are estimated, rather than an unconstrained ARMA. But \(f\left( \lambda \right) \) is naturally parameterised in terms of the zeroes of a and b, and when \(r\ge 2\) its form depends on the extent of any multiplicities in a,  whose number is likely unknown, even for given r. Note also that due to the aliasing property (6), a cannot be uniquely estimated from the AR coefficients of \(x_{t}\) when \(r\ge 2\), and though the presence of b may help, a Gaussian pseudo-likelihood may have multiple peaks from which it is difficult to choose (Robinson 1980b). Generally, corresponding to (6), there are uncountably many continuous processes consistent with a given discrete one, so attempting to estimate a continuous model is highly hazardous. It should be added that though (4) is classical, the model has two parts, the frequency response function  b / a and the noise \(\zeta \), and either might be specified differently, for example, the aliasing problem is avoided if \(\zeta \) is band limited to Nyqvist frequency, so instead of (5)

$$\begin{aligned} h\left( \lambda \right) =0,\left| \lambda \right| >\pi , \end{aligned}$$
(7)

leading to a quite different model for the discretely observed \(x_{t}\), see Robinson (1977a), while Robinson (1976a, b) evaluated bias of estimates based on (7) when in fact (4) holds, again using the theorem of residues. Extensions to vector \(x\left( t\right) \) were developed by Phillips (1974), Robinson (1993). The topic has been rediscovered more recently, with (4) referred to by the acronyms CAR (when \(s=0\)) and CARMA (when \(s>0)\).

For long memory, one can consider a continuous process such as

$$\begin{aligned} S\left( \lambda \right) =\frac{\lambda ^{-2d}}{2\pi }\frac{\left| b\left( i\lambda \right) \right| ^{2}}{\left| a\left( i\lambda \right) \right| ^{2}},\lambda \in \mathcal {R}, \end{aligned}$$

with memory parameter \(d\in \left( 0,1/2\right) \). Again (6) can be applied but there is now no neat analytic solution characterizing \(x_{t}\) and \(f\left( \lambda \right) \) (see Comte and Renault 1996; Chambers 1998).

For continuously defined spatial processes that are observed on a lattice, extensions of the above are clearly possible (see, e.g. Matsuda 2019), and analogous issues discussed for the time series case apply.

4 Miscellaneous topics

4.1 Nonstationary processes

Really, a time series AR with a unit root is a ’long memory’ process because it has even longer memory than a stationary long memory process, and it coincides with a fractional process with memory parameter \(d=1\). Asymptotic theory for unit root processes in an AR setting produces nonstandard limit distributiions, but nesting the unit root in the fractional, rather than AR, class leads to standard asymptotics, for example a central limit theorem with \(n^{1/2}\) norming and classical local power properties, for memory and other parameter estimates, even under nonstationarity, using either tapering and skipping of frequencies in discrete-frequency Whittle estimation (see Velasco and Robinson 2000), or using conditional sum-of-squares estimation (see Hualde and Robinson 2011). These ideas seem capable of extension to nonstationary fractional processes on a spatial lattice. Kuensch (1987) considered an alternative class of nonstationary spatial processes.

In a semiparametric time series setting, Velasco (1999) studied log-periodogram regression estimation under nonstationarity, using tapering and skipping of frequencies, as did Yajima and Matsuda (2019) for data on a spatial lattice of dimension \(M=2\) with an underlying continuous nonstationary isotropic process. In econometrics, study of nonstationary time series has much focussed on multivariate processes, and cointegration and fractional cointegration. Clearly, there is scope for corresponding study of multivarate spatial processes.

4.2 Irregular spacing

There is a time series literature on missing and irregularly spaced data, for example, Robinson (1977b) considered the underlying continuous model (4) with \(r=1\), \(s=0\), and showed that irregularly spaced observations have a kind of time-varying AR representation, with unknown parameters estimable by Gaussian pseudo likelihood. Irregular spacing is more likely to be an issue with spatial data. A general approach, for a parametric underlying continuous stationary spatial process, would be to construct a Gaussian pseudo-likelihood, conditional on the observation points. But conditions for asymptotic theory are messy relative to the regularly spaced case, mixing properties of the continuous process \(x\left( t\right) \) with those of the observation points. Matsuda and Yajima (2009) developed parametric and nonparametric methods and theory in the frequency demain, for spatial data with an underlying model defined on \(\mathcal {R}^{M}\). Another approach that could be pursued in an irregularly spaced long memory spatial setting involves averaging over the process generating the observation times as well as over \(x\left( t\right) \) itself. In particular, in the time series case, Robinson (1980b) considered the model (4) with irregularly spaced observation points \(\tau _{k}\), \(k\in \mathbb {Z}\), generated by a point process, deriving autocovariance properties that are not conditional on the observation times. For example, treating the time intervals as a renewal process, he obtained an ARMA(\(r,r-1\)) representation for the observed pseudo-regularly spaced process \(X_{k}=x\left( \tau _{k}\right) \), \( k\in \mathbb {Z}\).

Sometimes irregular spacing is due to missing from a regular lattice. Parzen (1963) proposed an amplitude modulating sequence for missing data in time series \(x_{t}\), \(t\in \mathbb {Z}\), studying the regularly spaced process \( a_{t}x_{t}\), where \(a_{t}\) is a (possibly stochastic) zero–one sequence, taking the value 1 when \(x_{t},\) is observed, and 0 when it is missed. Dunsmuir and Robinson (1981) studied Whittle estimation of parametric models using this approach. It seems clear how it could be extended to cover spatial lattice processes with missing observations and long memory.

4.3 Adaptive estimation in semiparametric models

As is well known, under regularity conditions Gaussian pseudo-maximum likelihood estimates for parametric models are typically consistent and asymptotically normal with the same asymptotic covariance matrix irrespective of whether or not Gaussianity actually holds. But such estimates are not asymptotically efficient if the proces is non-Gaussian. Considering a semiparametric model, with innovations having distribution of unknown form, it is possible to obtain asymptotically efficient ’adaptive’ estimates, using nonparametric smoothing. For long memory time series, this was studied by Hallin and Serroukh (1998) under stationarity, and by Robinson (2005) under stationarity and nonstationarity. Such ideas are certainly extendable to spatial lattice processes.

4.4 When ordering does not matter

Unlike with time series data, there is no natural uni-dimensional ordering of spatial data. For a spatial lattice, there is typically an ordering of locations only with respect to each of the dimensions. For inference on many features, such as spatial dependence parameters, respect for relative spatial locations is crucial. But for estimating some ’instantaneous’ features, such as location and static parametric regression, and in stochastic design nonparametric regression, ordering can be disregarded and the spatial process \(x_{i}\) \(i\in \mathbb {N},\) mapped arbitrarily into a sequence \(u_{i},\) \(i=1,\ldots ,n\).

We might assume, say (Robinson 2011), the nonstationary infinite MA

$$\begin{aligned} u_{i}=\sum \limits _{j=1}^{\infty }a_{ij}\eta _{j}, \ \ 1\le i\le n, \ \underset{i}{\sup }\sum \limits _{j=1}^{\infty }a_{ij}^{2}<\infty , \end{aligned}$$
(8)

which can be checked in some spatial model settings. We have ’long memory’ if

$$\begin{aligned} \frac{1}{n}\sum \limits _{i,j=1:i\ne j}\mathrm{Cov}(u_{i},u_{j})= \frac{1}{n}\sum \limits _{i,j=1:i\ne j}^{n}\sum \limits _{k=1}^{\infty } a_{ik}a_{jk}\rightarrow \infty ,\text { as } n\rightarrow \infty . \end{aligned}$$

For stochastic design nonparametric (kernel) regression, with spatial data, Robinson (2011) gave conditions for asymptotic normality, based partly on assuming (8) for the error process and on conditions on the distance between multivariate and product univariate densities for the regressors, which can also cover long memory. Such conditions might be extended to other models and methods for spatial data.