Abstract
Shifted convolution sums play a prominent rôle in analytic number theory. Here these sums are considered in the context of holomorphic Hecke eigenforms. We investigate pointwise bounds, mean-square bounds consistent with the optimal conjectural bound, and find asymptotics on average for their variance.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Sums of the form
play an important rôle in analytic number theory, especially when \(\lambda \) is of arithmetic significance, see e.g. [2, 4, 6, 11, 14, 16, 18, 37] and the references therein. The case where \(\lambda (n)\) is the nth Hecke eigenvalue of an automorphic object is maybe the most interesting, and in this case the above sum is called a shifted convolution sum or sometimes a generalized additive divisor sum. Here are some examples:
-
(a)
When \(\lambda (n)\) equals the divisor function \(d(n)=\sum _{b\vert n} 1\) this is the classical binary additive divisor problem. In this case \(\lambda (n)\) are the Hecke eigenvalues (of the derivative in s at \(s=1/2\)) of the weight 0 non-holomorphic Eisenstein series E(z, s) for the full modular group. See e.g. [20, p. 62].
-
(b)
When \(\lambda (n)\) equals \(r_2(n)\) i.e. the number of ways of representing n as the sum of two squares and \(h=1\) this is a special case of the hyperbolic circle problem, see e.g. [20, p. 174]. In this case \(\lambda (n)\) is equal to the Hecke eigenvalues of the theta series \(\theta _2(z)=\sum _{n\in \mathbb {Z}^2} e^{2\pi i \left\Vert n \right\Vert ^2 z}\).
In both of the above cases the automorphic object is non-cuspidal and the sum A(X, h) admits an asymptotic expansion with a main term \(XP(\log X)\) where P is a polynomial, and an error term which is provably \(O_h(X^{2/3+{\varepsilon }})\) (See [6, Thm. 1][29, Cor. 1 ], [20, Chapter 12]), and conjecturally \(O(X^{1/2+{\varepsilon }}h^{\varepsilon })\).
In this paper we consider the automorphic object to be a weight k, level 1 holomorphic cuspidal Hecke eigenform \(f\in S_k(1)\), normalized such that its Fourier expansion
satisfies \(\lambda _f(1)=1\). Here \(e(z)=e^{2\pi i z} \). We note that, since the Hecke operators are self-adjoint, \(\lambda _f(n)\) is real. In this cuspidal case there is no main term but the provable error term is still of the same size as in the non-cuspidal case, i.e.
uniformly for \(1\le h\le X^{2/3}\). This was proved by Jutila [21, Eq. (1.32)], and we sketch a variant of his proof in Sect. 2. Note that Jutila uses a different normalization on \(\lambda _f(n)\). Also in this case we expect the conjectural bound
for \(h\le X^{1/2-{\varepsilon }}\). Note that the implied constant depends on f. Probably this implied constant is bounded by a constant times \(k^{\varepsilon }\) at least for certain ranges of k vs X, but the precise conjectural range is not obvious.
One frequently encounters smooth sums e.g.
where \(W:{\mathbb R}_+\rightarrow {\mathbb R}\) is smooth and supported in a compact interval, e.g. on [1, 2]. Note that we are summing over the range where the average of n and \(n+h\) is of size roughly X. It is perhaps not too surprising that the analogue of the conjectural bound (2) holds for the smooth sum \( A_f^{W}(X,h)\), and we will give a short proof of this well-known fact (see [2, Eq. (5)]).
Proposition 1.1
Assume the Ramanujan–Petersson conjecture for Maass forms. Then
for \(h=O(X^{1/2-{\varepsilon }})\).
Without assuming the Ramanujan–Petersson conjecture one gets a slightly weaker bound, but one which holds in a larger range of h. See Proposition 2.2 below for details, as well as comparison with similar results in the existing literature.
For non-smooth sums we prove the same bound in the mean-square.
Theorem 1.2
For \(h\le X^{1/2}\) we have
Note that for fixed f, h this is consistent with the conjectural bound of \(A_f(x,h)\) and for varying h it is better than the trivial bound coming from (21) as long as \(h\le X^{1/2}\). The result is analogous to a result proved by Faĭziev [8, Theorem 3] and independently by Ivić and Motohashi [17, Cor. 5], who proved the analogous result for the classical additive divisor problem. For the case (b), i.e. \(\lambda (n)=r_2(n)\) the analogous result is due to Chamizo [3].
The main ingredient in proving Proposition 1.1 and Theorem 1.2 is the spectral expansion of the relevant generating series.
When we average \(A_f^{W_1}(X,h_1)\overline{A_f^{W_2}(X,h_2)}\) over a Hecke basis \(H_k\) of \(S_k\) (with suitable weights) we find asymptotics as \(X,k\rightarrow \infty \) at a certain range. These are consistent with \(A_f^{W}(X,h)\) being of the order of \(X^{1/2}\) in the X-aspect when \(X\ll k^{1/2-{\varepsilon }}\). More precisely we have the following:
Let \(\tau _1(n)=\sum _{d\mid n} d\) be the sum of divisors function, and let \(L(s, {\text {sym}}^2 f)\) be the symmetric square L-function associated to f, i.e.
and defined on \({\mathbb C}\) by analytic continuation. For \(W_1, W_2 :(0,\infty )\rightarrow {\mathbb R}\) and \(h_1,h_1\in {\mathbb N}\) we let
Then we show the following result:
Theorem 1.3
Let \(W_1, W_2 :(0,\infty )\rightarrow {\mathbb R}\) be smooth with compact support. Then for \(X\ll k^{1/2-{\varepsilon }}\) we have
We recall that the arithmetical weights \(L(1, {\text {sym}}^2 f)^{-1}\) are relatively well-behaved, in the sense that \(k^{-{\varepsilon }}\ll L(1, {\text {sym}}^2 f)\ll k^{\varepsilon }\), see [9]. With some work these weights can probably be removed from the Theorem. The interest in Theorem 1.3 comes from the study of small scale equidistribution at infinity for Hecke eigenforms, see [32], where sums with extra average over k are considered. These results are proved using the Petersson formula. In the case of Theorem 1.3 the off-diagonal terms arising from the formula are essentially trivial, due to the decay of the J-Bessel function.
2 Spectral expansions of shifted convolution sums
In order to understand the average behavior of \(\lambda _f(n)\lambda _f(n+h)\) we consider, following Selberg and others e.g. [12, 21, 25, 35, 37], the Dirichlet series
This series converges absolutely for \(\Re (s)>1\) by standard bounds on the Fourier coefficients, see, e.g. Deligne’s bound [5]
or the Rankin–Selberg estimate [34, 36]
We note that since the Fourier coefficients are real,
In order to understand the analytic properties of \(D_f(s,h)\) we consider, following Selberg [37, (3.10)], the Poincaré series
where \(h\in {\mathbb N}.\) This is convergent for \(\Re (s)>1\) and admits meromorphic continuation on \(s\in {\mathbb C}\) as a \(\Gamma \)-automorphic function. In \(\Re (s)\ge 1/2\) its poles are at \(s_j=1/2+it_j\), where \(1/4+t_j^2\) is an eigenvalue of the automorphic Laplacian, and in \(\Re (s)>1/2\) it is square integrable away from its poles. Notice that for \(\Gamma =\Gamma (1)\) there are no poles for \(\Re (s)>1/2\). By unfolding we find that
Since the left-hand side is meromorphic for \(s\in {\mathbb C}\), this gives the meromorphic continuation of \(D_f(s,h)\) to \(s\in {\mathbb C}\). The function \(U_h(z,s)\) is orthogonal to constants by unfolding, and has the spectral expansion
where \(\{u_j\}_{j=1}^\infty \) is an orthonormal basis of Maass cusp forms and E(z, s) is the non-holomorphic Eisenstein series at the cusp \(i\infty \), see [20, Thms 4.7 and 7.3]. Again by unfolding we have
where
is the Fourier expansion of \(u_j\) at infinity and, similarly,
is the nth Fourier coefficient of E(z, s).
The expansion (7) combined with (6) gives the following expression, which is useful for understanding the analytic properties of \(D_f(s,h)\), cf. [13, Theorem 1] and [21, (1.20)]:
where
and
In order to effectively analyze \(D_f(s,h)\) using the spectral expansion (8) we need the following bound due to Good [10, Theorem 1]:
See also [21, Theorem 1] for the case of a Maass form \(u_{j_0}\) instead of \(y^{k/2}f\). Comparing with Weyl’s law [20, Corollary 11.1], we observe that (9) implies that on average \(c_j(f)\) is bounded by a constant times \(e^{-\pi t_j/2}t_j^{k-1}\). Note that, by using a triple product identity, this may be interpreted as the Lindelöf hypothesis on average for \(L(f\times f\times u_j,1/2)\) in the spectral aspect, see [15, 38].
Concerning the Fourier coefficients we have for \(\Gamma ={\hbox {SL}_2( {\mathbb Z})} \) the following bound, which follows from Kuznetsov’s asymptotic formula [24, Theorem 6], see also [30, Lemma 2.4] and the remarks following its proof:
uniform in h, T and \(1\le \Delta \le T\). Note that Kuznetsov’s result is stated without the integral but a direct investigation of the integral gives \(\ll T(Th)^{\varepsilon }\). We note that again here we are using the arithmeticity of \(\Gamma \). For a more general group the best known bound is \(T^2+hT\) in a long range \(\Delta =T\), see [20, Eq (9.13)]. Using these (9) and (10) we can prove the following result:
Theorem 2.1
The function \(D_f(s,h)\) admits meromorphic continuation to \(s\in {\mathbb C}\). On \(\Re (s)\ge 1/2\) the poles are exactly at \(s=1/2\pm it_j\), with corresponding residues
For \(1/2+{\varepsilon }\le \Re (s)\le 1+{\varepsilon }\) we have the bound
Proof
The holomorphic continuation to \(\Re (s)>1/2\) follows from the spectral expansion (8), as the sum and integral converge absolutely and locally uniformly when \(\Re (s)\ge 1/2+{\varepsilon }\). This follows from the following bound, which also proves the claim about the bound on vertical lines: Using Cauchy–Schwarz in combination with (9) and (10) we find that for U bounded away from zero we have
Note that for s in a vertical strip, say \(s=\sigma +i\tau \) with \(-\infty<a\le \sigma \le b<\infty \), when all arguments of the Gamma functions are bounded away from \(0, -1, -2, \ldots \), the Stirling asymptotics gives
It is convenient to note that the bound is invariant under changing signs on \(\tau \) as well as on t. The exponential factor is trivial for \(\left|t \right|<\left|\tau \right|\) and equals \(e^{-\pi (\left|t \right|-\left|\tau \right|)}\) otherwise. We split the sum and integral in (8) according to the various ranges of the spectral parameter t. We assume \(\left|\tau \right|>1\).
-
(i)
If the spectral parameter satisfies \(\left|\left|t \right|-\left|\tau \right| \right|\le 1\), then
$$\begin{aligned} G_k(s)\ll (1+\left|s \right|)^{-(k+\sigma -1)}, \end{aligned}$$so combining this with (11) we find that both the corresponding discrete (\(\left|t_j-\left|\tau \right| \right|\le 1\)) and continuous (\(\left|\left|t \right|-\left|\tau \right| \right|\le 1\)) contributions are bounded by
$$\begin{aligned} \left|s \right|^{1-\sigma }(\left|s \right|^{1/2}+h^{1/4})(h\left|s \right|)^{\varepsilon }. \end{aligned}$$ -
(ii)
If the spectral parameter satisfies \(\left|\left|t \right|-\left|\tau \right| \right|> 1\) with \(0\le \left|t \right|\le 2\left|\tau \right|\), then
$$\begin{aligned} G_k(s,t)\ll (1+\left|\left|\tau \right|-\left|t \right| \right|)^{\sigma -1}(1+\left|s \right|)^{-(k+\sigma -1)}. \end{aligned}$$We further subdivide the sum into \(1\le A<\left|\left|t \right|-\left|\tau \right| \right|\le 2A\) with \(A \le \left|\tau \right|\). Note that this is a sum at height \(\ll \left|\tau \right|\) of length \(\ll A\) and we find that the sum over this range is bounded by a constant times \((1+A)^{\sigma -1}(1+\left|s \right|)^{-(\sigma -1)}(\tau ^{1/2}A^{1/2}+h^{1/4})\left|sh \right|^{\varepsilon }\). We therefore bound the contributions as
$$\begin{aligned} \sum _{\begin{array}{c} 0\le t_j\le 2\left|\tau \right|\\ \left|t_j-\left|\tau \right| \right|> 1 \end{array}}&\ll \sum _{1\le n \le \log _2(\left|\tau \right|)}\sum _{2^n\le \left|t_j-\left|\tau \right| \right|\le 2^{n+1}}\\&\ll \sum _{1\le n \le \log _2(\left|\tau \right|)}\frac{(2^n)^{\sigma -1}}{(1+\left|s \right|)^{(k+\sigma -1)}}(\tau ^{1/2}2^{n/2}+h^{1/4})\left|sh \right|^{\varepsilon }\\&\ll \frac{1}{(1+\left|s \right|)^{(\sigma -1)}}(\left|\tau \right|^{1/2}\max (1, \left|\tau \right|^{\sigma -1/2})+h^{1/4})\left|sh \right|^{\varepsilon }\end{aligned}$$and similarly for the continuous contribution. When \(\sigma >1/2\), this agrees with the bound stated in the theorem.
-
(iii)
If \(n\left|\tau \right|< \left|t \right|\le (n+1)\left|\tau \right|\) with \(n\ge 2\) (i.e. a window of size \(\left|\tau \right|\) at height \(n\left|\tau \right|\), we see that
$$G(s,t)\ll e^{-\pi (n-1)\left|\tau \right|}\frac{n^{2\sigma -2}}{(1+\left|s \right|)^{k}},$$and we bound the contribution as
$$\begin{aligned} \sum _{n\left|\tau \right|< t_j\le (n+1)\left|\tau \right|} \ll e^{-\pi (n-1)\left|\tau \right|} n^{2\sigma -2}((n\left|\tau \right|^2)^{1/2}+h^{1/4})\left|hs \right|^{{\varepsilon }}. \end{aligned}$$We can sum these contributions over n from 2 to infinity and get a bound of \(\ll h^{1/4}\left|s \right|^{-A}\) for any A.
For \(\left|\tau \right| < 1\) the function G(s, t) decays exponentially in t locally uniformly in s and the sum and integral are again holomorphic.
To see the meromorphic continuation further to the left we argue as follows: The sum in (8) is meromorphic in \({\mathbb C}\) with poles at \(s=1/2\pm i t_j-l\), \(l=0,1,\ldots \) with corresponding residues
We note that for \(1/2-l+{\varepsilon }\le \Re (s)\le 1/2-l+1-{\varepsilon }\) the above bound on vertical lines is still valid for the sum with the same proof as long as we are bounded away from the poles.
The integral in (8) can be written as a line integral along the vertical line \(\Re (w)=1/2\):
This can be continued analytically to \(s\in {\mathbb C}\) as follows: By holomorphicity we may deform the integral to the left of two small discs centered at \(1/2\pm i\tau \). This continues the integral to a holomorphic function in the union of \(\Re (s)>1/2\) with the two small discs. The integral (12) is also holomorphic for \(-1/2<\Re (s)<1/2 \) by absolute convergence using the above estimates. For this expression we can now deform the line of integration to the right of two small half-discs at the same heights \(\pm \tau \). These two holomorphic functions are now both defined in the two discs. To see how these two functions relate, we take the difference of the functions and find, by Cauchy’s residue theorem, that their difference equals
It follows that the integral
can be continued to a meromorphic function in \(\Re (s)>-1/2\) that is analytic in \(\Re (s)\ge 1/2\) and for \(-1/2<\Re (s)<1/2\) can be expressed by
We may repeat this process to get continuation to \(s\in {\mathbb C}.\) \(\square \)
Remark 1
We note that
which makes the residues in Theorem 2.1 explicit. We note also that the proof gives also the location of all the poles of \(D_f(s,h)\) and allows us to compute the residues at all of them including at the residues of Eisenstein series.
Concerning the bound on \(D_f(s,h)\) it can be improved as follows. Assume that the Satake parameters for \(u_j\), say \(\alpha _j(p)\), \(\beta _j(p)\) satisfy
We call this bound \(H(\theta )\). By the Hecke relations for Hecke eigenvalues it follows that the Hecke eigenvalues for \(u_j\) satisfy \(\left|\lambda _j(n) \right|\le d(n)n^\theta \). By the work of Kim and Sarnak [23] H(7/64) is true, and the Ramanujan–Petersson conjecture predicts that H(0) is true. Assuming \(H(\theta )\) we find that the left-hand side of (10) is \(O(h^{2\theta }T\Delta (hT)^{\varepsilon })\). Using this in the proof of Theorem 2.1 we find that for \(1/2+{\varepsilon }\le \sigma \le 1+{\varepsilon }\) we have the bound
Since for \(\sigma \ge 1+{\varepsilon }\) we have \(D_f(s,h)\ll _{\varepsilon }1\), by (4) or (5), the Phragmén–Lindelöf principle gives for \(1/2+{\varepsilon }\le \sigma \le 1+{\varepsilon }\) the estimate
Remark 2
Note that the bound in Theorem 2.1 and its proof is very similar in spirit to [35, Thm A.1]. The reason we get a better bound than Sarnak is that we move the line of integration further to the left (there are no small eigenvalues for the full modular group) and we use Good’s average bound (9); Sarnak is using individual bounds [35, (A18)], which are not as strong. A similar analysis is implicit in [21, p. 459].
From Theorem 2.1 combined with the bound in Remark 1 we can now prove the following bound on the smooth sum defined in (3):
Proposition 2.2
Assume that \(H(\theta )\) is true for some \(\theta <1/2\). Then
for \(h=O(X^{\frac{1}{2(1-\theta )}})\).
Proof
By a standard complex contour argument using Mellin inversion, see e.g. [31, Appendix A], we find
Using the mean value theorem we easily find that
Combining this with (4) or (5) we see that the left-hand side of (14) equals
from which the claim follows easily. \(\square \)
Remark 3
Proposition 2.2 should be compared with Blomer’s bound [1, Thm 1.3] as well as Blomer and Harcos’s bound [2, Eq. (5)]. These results prove a similar bound. Blomer uses Jutila’s variant of the circle method combined with the spectral large sieve inequalities of Deshouillers and Iwaniec [7], whereas Blomer and Harcos use the spectral theory on \(\hbox {GL}_2\) rather than \(\hbox {SL}_2\). The results of Blomer and Harcos provides bounds which are more uniform in the various parameters. The above argument serves as a simpler proof, which recovers the same strength in the X-parameter. Note that for the divisor function and for \(r_2(n)\) (case (a) and case (b) in the introduction) we have \(\Omega \) results in X for the error term, see [18] and [33]. These are, up to logarithms, of the same order as the upper bound in Proposition 2.2 as X tends to infinity.
In order to get good estimates on non-smooth sums we use Jutila’s explicit formula, which we now explain. For this it is convenient to use a slightly modified sum namely
Note that by Deligne’s bound (4) and the trivial estimate (15) we have for \(h\le X\)
For \(l=0,1,2,\ldots \) and \(t\ne 0\) we define
and note that for \(\left|t \right|\) bounded away from zero we have, by Stirling’s formula,
Proposition 2.3
Assume that \(1\le h, T\le X\), and \(Th\le X^{1-\delta }\) for some \(\delta >0\). Then there exists a \(B>0\) such that
A similar result is stated in [21, (1.29)]. For the reader’s convenience we sketch a proof:
Proof
Perron’s formula, see e.g. [28, Cor. 5.3], gives for \(T\gg 1\)
where
Here we have used that
as follows from the Rankin–Selberg bound (5) or from Deligne’s bound (4). Using Deligne’s bound once more, we can bound the sum in R and find \(R=O(({X}/{T})X^{\varepsilon })\).
We now move the line of integration to \(\Re (s)=1/2+{\varepsilon }\). The contribution from the corresponding horizontal lines can be analyzed as follows. Using Theorem 2.1 and the estimate \(D_f(s,h)\ll _{\varepsilon }1\) for \(\Re (s)=1+{\varepsilon }\), we note that on both horizontal lines \(\Re (s)=1/2+{\varepsilon }, 1+{\varepsilon }\) we have
Let \(g(s)=\left( Xs^{-1}+X^{1/2}h^{1/4}s^{-1/2}+X^{1/2}\right) (Xs)^{\varepsilon }\), where we have chosen the principal powers. It is elementary to verify that for \(\Re (s)>{\varepsilon }\) we have
We therefore have \(D_f(s,h)/g(s)\ll 1\) on both lines \(\Re (s)=1/2+{\varepsilon }, 1+{\varepsilon }\) with a constant not depending on h, X and we may conclude from the Phragmén–Lindelöf principle that (18) holds for all \(1/2+{\varepsilon }\le \Re (s)\le 1+{\varepsilon }\). It follows that the contributions of the horizontal lines in the integral are bounded by a constant times
Note that since \(h\le X\) the middle term is bounded by \(X^{3/4}/T^{1/2}\), which is always smaller than one of the two remaining terms, so it may be ignored.
We now define
with
For s with \(\Re (s)=1/2+{\varepsilon }\) and \(\left|\Im {s} \right|\le T\) we bound \(D_f(s,h)-S_T(s,h)\) as in the proof of Theorem 2.1. Note that in all remaining contributions coming from (8) the bound on \(G_k(s,t)\) has a factor of \(e^{-\pi (\left|t \right|-\left|\tau \right|)}\le e^{-\pi (T-\left|\Im (s) \right|)}e^{-\pi /2(\left|t \right|-\left|\tau \right|)}\). We may pull out the factor \(e^{-\pi (T-\left|\Im (s) \right|)}\) and bound the rest exactly as in Theorem 2.1 with the same result and we find under these conditions that
Inserting this we find
We now again bound all terms in \(S_T(s,h)\) exactly as in the proof of Theorem 2.1 and find that on \(\Re (s)=1/2+{\varepsilon }\) we have
We can therefore extend the s-integral to the line from \(1/2+{\varepsilon }-i(T+1)\) to \(1/2+{\varepsilon }+i(T+1)\) at no additional cost; this is useful in order to keep the distance from the poles when we move the line of integration further to the left.
We now analyze the integral in
by analyzing the contributions from \(S_T^d(s,h)\) and \(S_T^c(s,h)\) separately.
To analyze the contribution coming from \(S_T^d\) we move the line of integration to \(\Re (s)=-B\) picking up poles of \(S_T^d(s,h)\) at the poles of the Gamma functions and at 0. The poles of the Gamma function contribute exactly the sum in the theorem as can be seen from (13), so we need to show that for a suitable choice of B all other contributions give admissible errors.
Note that \(G_k(t,0)=0\) so \(S_T^d(0,h)=0\) and 0 is a removable pole of the integrand. We note that by the proof of Theorem 2.1 we have
We recall that all terms were in absolute value, so a short sum is bounded by a longer one. This implies that the contribution from the horizontal lines can be bounded by
which is an admissible error. The contribution from the line \(\Re (s)=-B\) is bounded by
Since \(Th\le X^{1-\delta }\), the constant B may be chosen large enough such that this is also an admissible error. This finishes the analysis of the discrete part of the spectrum.
To analyze the contribution from the continuous spectrum \(S_T^c(s)\) we need, after changing the order of integration, to estimate
After moving the line of integration to \(-B'\) the inner integral equals
Bounding using Stirling’s formula on all Gamma factors we find that this is bounded by a constant times
Inserting that back in (19) we see that for \(B'\) large enough the contribution from this part of the u-integral is \(O(X^{1/2+{\varepsilon }})\). We conclude that the contribution of the continuous spectrum is \(O(X^{1/2+{\varepsilon }})\). \(\square \)
We notice that by (17) the factors in the spectral expansion in Proposition 2.3 for \(\left|t \right|\) bounded away from zero can be bounded as
Bounding everything trivially, i.e. using (9) and (10), we find that
We choose \(T={X^{1/3-{\varepsilon }}}\) to balance the error terms. This gives
uniformly for \(h\ll X^{2/3}\). Combined with (16) this recovers (1).
2.1 Mean square bounds
In order to prove Theorem 1.2 we first state a general result loosely stating that the mean square of a function with certain properties is small if and only the function only has few well-spaced points at which the function is large. The proof is a relatively straightforward adaptation of [17, Sect. 4].
Lemma 2.4
Consider a measurable function \(\left|B(X,h) \right|\) defined for \(X\ge 2\). Assume that \(\left|B(X,h) \right|\le X^A\) for some \(A>0\) independent of h, and that for \(1\le Y\le X\)
uniformly for h in some set \(A_{X}\). Let \(C_h>0\) be a constant depending on h. Then the following are equivalent:
-
(1)
For every \({\varepsilon }>0\) we have
$$\begin{aligned} \int _{X}^{2X}\left|B(t,h) \right|^2dt\ll _{\varepsilon }C_hX^{2+{\varepsilon }} \end{aligned}$$uniformly in \(h\in A_X\).
-
(2)
Given \(X\le X_1<X_2<\cdots <X_{R}\le 2X\) with spacings \(X_{r+1}-X_r\ge V\ge X^{1/2+{\varepsilon }}\) and satisfying \(\left|B(X_r,h) \right|\ge V\). Then
$$\begin{aligned}R\ll _{\varepsilon }C_h X^{2+{\varepsilon }}V^{-3} \end{aligned}$$uniformly in \(h\in A_X\).
To see that we may apply Lemma 2.4 to \(B_f(X,h)\) we note that by Deligne’s bound (4) we have
uniformly in h. Also using Deligne’s bound we have for any with \(0\le t\le X\)
uniformly in h. Integrating over t we find that for \(1\le Y\le X\)
and using Cauchy–Schwarz we find
We conclude that the assumptions of Lemma 2.4 are satisfied.
We now verify that Lemma 2.4 (2) holds: Let \(X\le X_1<X_2<\cdots <X_{R}\le 2X\) with spacing \(X_{r+1}-X_r\ge V\ge X^{1/2+{\varepsilon }}\) and satisfying \(\left|B_f(X_r,h) \right|\ge V\). We may assume, without loss of generality, that
for if \(V\ge X^{2/3+{\varepsilon }}\) it follows from (21) that for X sufficiently large there does not exist any \(X\le X_r\le 2 X\) satisfying \(\left|B_f(X_r,h) \right|\ge V\) and in this case the condition is clear.
We assume \(R\ge 1\). For each r we choose a complex number \(\theta _r\) with norm 1 such that \(\theta _r B_f(X_r,h)\ge V\). Now Proposition 2.3 gives that when \(hT\le X^{1-\delta }\) we have
Using dyadic decomposition, Cauchy–Schwarz, (20), and (9), the sum can be bounded by
where the index \(t_j\sim T\) means that we are summing over \(t_j\) satisfying \(T\le t_j\le 2T\). In order to estimate this we use a large sieve inequality due to Jutila [22]:
Theorem 2.5
Let \(X\le X_1<X_2<\cdots <X_{R}\le 2X\) with spacing \(X_{r+1}-X_r\ge V>0\), and let \(1\le T\), \(1\le \Delta \le T\) and \(1\le N\). Then for any \({\varepsilon }>0\) we have
where \(b=(b_{n,r})\) is any complex vector and \(\left\Vert b \right\Vert \) is the standard \(l^2\)-norm of b.
Proof
This is [22, Theorem 4.1] specialized to \(\varphi _n(x,y)\)=1, \(\psi _n(x,y)=\pm x\log (1+y)/(2\pi )\), \(\Psi =\Psi '=T\), \(\lambda =\Phi =1\), \(y_r=X_r/X-1\), \(\delta =V/X\). \(\square \)
Applying Theorem 2.5 we find that the sum in (22) is bounded by an absolute constant \(C_{\varepsilon }\) times
where we are using that \(Th/X\ll 1\). This gives
We choose \(T=X^{1+{\varepsilon }/2}/V\) and note that with this choice the last term satisfies
since \(V\ge X^{1/2+{\varepsilon }}\). It follows that
Squaring and dividing by \(RV^2\) we find that
If \(h\le X^{1/2}\) we have \(hT/X\le {X^{1/2+{\varepsilon }/2}}/{V}\le X^{-{\varepsilon }/2}\) so all estimates hold uniformly in this range; this requirement comes from \(Th\le X^{1-\delta }\) in Proposition 2.3. Using Lemma 2.4 we conclude the following result:
Theorem 2.6
Let \(h\ge 1\). Then
uniformly for \(h\ll X^{1/2}\).
Using (16) we see that this also implies Theorem 1.2. We note that in order to improve on the h dependence with this method it is crucial to improve on the factor \((T+N)\) in Theorem 2.5.
3 The variance over a Hecke basis
We now go back to the smooth shifted convolution sums \(A_f^{W}(X,h)\), and investigate the variance over an orthonormal basis of Hecke eigenforms \(H_k\). More precisely we want to understand
In order to better describe the dependence on W, and h we use the Sobolev norms
For all compactly supported functions W we assume \(\mathrm{supp\,} {W_i}\subseteq [a_W,A_W]\), and that \(A_W\ge 1\), and for two shifts \(h_1\), \(h_2\) we denote \(\left\Vert h \right\Vert _\infty =\max (h_1,h_2)\).
3.1 Small range asymptotics
We investigate the case where the size of the range of the sum in \(A_f^{W}(X,h)\) is small compared to the size of the additional average i.e. \(\dim (S_k)\sim k/12\). Specifically we investigate the range \(X\le k^{1/2-{\varepsilon }}\).
The main tool in understanding this is the Petersson formula, which states that
see e.g. [27, p. 776]. When \(X\le k^{1/2-{\varepsilon }}\) we will see that we can make good use of the following bound on the J-Bessel function
see e.g. [26, p. 233]. In fact the decay (24) will imply that the contribution from the sum over c becomes essentially neglegible for this problem. This means that we are reduced to studying the diagonal terms. In our treatment we take extra care to get an explicit error term.
In order to use the Petersson formula we note that, since the normalized Hecke eigenvalues satisfy the Hecke relations
see [19, (6.38)], we may rewrite
To state our theorem in this case we recall from the introduction that for functions \(W_1, W_2:(0,\infty )\rightarrow {\mathbb R}\) and \(h_1,h_2\in {\mathbb N}\) we define
Theorem 3.1
Let \(W_1, W_2:(0,\infty )\rightarrow {\mathbb R}\) be smooth functions with compact support. Then for \(A_{W_1}A_{W_2}X\ll k^{1/2-{\varepsilon }}\), \(h_i<2a_{W_i}X\),
The implied constant is \(\ll _{\varepsilon }\left\Vert h \right\Vert _\infty ^{{\varepsilon }}(\left\Vert W_1(h_1\cdot )W_2(h_2\cdot ) \right\Vert _{2,1} +\left\Vert W_1 \right\Vert _\infty \left\Vert W_2 \right\Vert _\infty )\) for any \({\varepsilon }>0\).
Proof
Using (25) and the Petersson formula (23) we find
We refer to the line with the Kronecker delta as the diagonal term, and the rest as the off-diagonal term.
To handle the diagonal term, we observe that for fixed positive \(d_1\ne d_2\) the equation
has only finitely many positive integer solution \((r_1,r_2)\in {\mathbb N}^2\). To see this we rewrite the equation as \((2r_1+d_1)^2-(2r_2+d_2)^2=d_1^2-d_2^2\). Factoring the left-hand side as \((2r_1+d_1+2r_2+d_2)(2r_1+d_1-2r_2-d_2)\) we see that any solution gives a factorization of \(d_1^2-d_2^2\), and that any factorization of \(d_1^2-d_2^2\) comes from at most one solution. This shows that there are at most \(d(d_1^2-d_2^2)\) solutions to (27) with \(d_1\ne d_2\), where d(n) denotes the number of divisors of n; indeed we see that the total contribution from these terms is \(O(\left\Vert h \right\Vert _\infty ^{ {\varepsilon }}\left\Vert W_1 \right\Vert _\infty \left\Vert W_2 \right\Vert _\infty )\).
For the remaining terms, i.e. \(d_1=d_2=d\), \(r_1=r_2=r\) we apply first Poisson summation in the r-variable and use the fact that the Fourier transform of the function \(y\mapsto W_1\left( \frac{h_1(y+d/2)}{dX} \right) \overline{W_2\left( \frac{h_2(y+ d/2)}{dX}\right) }\) at r is bounded by an absolute constant times
This follows from repeated integration by parts.
By the assumption on \(h_i\) we now see that
where, in the first equality we have extended the r-sum to all of r trivially as all added terms are zero, since \(h_i<2a_{W_i}X\). In the second equality we use Poisson summation and finally evaluate the \(r=0\) term and bound the rest using (28) with \(n=2\).
Using the bound (24) on the J-Bessel function, we may bound the off-diagonal term in the expression for (26) by an absolute constant times \(\left\Vert W_1 \right\Vert _\infty \left\Vert W_2 \right\Vert _\infty \) times
using the trivial bound for Kloosterman sums and the sup norm for the test functions. Note that since \(r_i(r_i+d_i)=(r_i+d_i/2)^2-d_i^2/4\) the square root is \(\ll A_{W_1}A_{W_2}X^2\). Since we have assumed that \(A_{W_i}X\le A_{W_1}A_{W_2} X\ll k^{1/2-{\varepsilon }}\), we find
and, therefore, the off-diagonal term decay exponentially in k and only polynomially in X. This finishes the proof, as it shows that the contribution of the off-diagonal term is \(O(\left\Vert h \right\Vert _\infty ^{\varepsilon }\left\Vert W_1 \right\Vert _\infty \left\Vert W_2 \right\Vert _\infty k^{-A})\) for any \(A>0\). \(\square \)
Remark 4
Note that in Theorem 3.1 the leading term is coming from the diagonal term and the off-diagonal term is essentially negligible because of the exponential decay of the J-Bessel function when \(X\ll k^{1/2-{\varepsilon }}\). If \(X\gg k^{1/2-{\varepsilon }}\) such decay is not available, and we do not know how to estimate the sum. However, if we make an additional average over k, then the off-diagonal may be analyzed, and it is again possible to prove an asymptotic formula. See [32, Thm 2.1] for details.
Remark 5
We can remove the condition \(h_i<2a_{W_i}X\) in Theorem 3.1 as follows. The only place this restriction is used is to extend the summation in (29) to \(\mathbb {Z}\) instead of \({\mathbb N}\) since in this case there are no \(r\le 0\) with \(a_{W_i}({d_i}/{h_i})X\le r_i+ {d_i}/{2}\). Without \(h_i<2a_{W_i}X\) there can be at most \(d_i\) such non-positive r as \(a_{W_i}>0\) so we may omit the restriction on \(h_i\) at the expense of an extra error term of size \(O(\left\Vert h \right\Vert _\infty ^{1+{\varepsilon }}\left\Vert W_1 \right\Vert _\infty \left\Vert W_2 \right\Vert _\infty \)).
Remark 6
We can also prove Theorem 3.1 when \(W_i(y)=1_{y\le 1}\) are sharp cut-off functions, and we get an error term \(O(\left\Vert h \right\Vert _\infty ^{1+{\varepsilon }})\). In this case the Poisson summation argument is replaced by an elementary count. An interpretation of Theorem 3.1 with sharp cut-offs is that we obtain the conjectured square-root cancellation in the X-aspect for the shifted convolution problem on average.
Data availability
Data sharing not applicable to this article as no datasets were generated or analysed during the current study.
References
Blomer, V.: Shifted convolution sums and subconvexity bounds for automorphic L-functions. Int. Math. Res. Not. 73, 3905–3926 (2004)
Blomer, V., Harcos, G.: The spectral decomposition of shifted convolution sums. Duke Math. J. 144(2), 321–339 (2008)
Chamizo, F.: Some applications of large sieve in Riemann surfaces. Acta Arith. 77(4), 315–337 (1996)
Conrey, J.B., Gonek, S.M.: High moments of the Riemann zeta-function. Duke Math. J. 107(3), 577–604 (2001)
Deligne, P.: La conjecture de Weil. I. Inst. Hautes Études Sci. Publ. Math. 43, 273–307 (1974)
Deshouillers, J.-M., Iwaniec, H.: An additive divisor problem. J. Lond. Math. Soc. Second Ser. 26(1), 1–14 (1982)
Deshouillers, J.-M., Iwaniec, H.: Kloosterman sums and Fourier coefficients of cusp forms. Invent. Math. 70(2), 219–288 (1982)
Faĭziev, R.F.: Estimates in the mean in the additive divisor problem, Izvestiya Akademii Nauk Tadzhikskoĭ SSR. Otdelenie Fiziko-Matematicheskikh, Khimicheskikh i Geologicheskikh Nauk 1(95), 7–17 (1985)
Goldfeld, D., Hoffstein, J., Lieman, D.: An effective zero-free region. Ann. Math. 140(1), 177–181 (1994)
Good, A.: Cusp forms and eigenfunctions of the Laplacian. Math. Ann. 255(4), 523–548 (1981)
Good, A.: On various means involving the Fourier coefficients of cusp forms. Math. Z. 183(1), 95–129 (1983)
Good, A.: The square mean of Dirichlet series associated with cusp forms. Mathematika 29(2), 278–295 (1983)
Hafner, J.L.: Explicit estimates in the arithmetic theory of cusp forms and Poincaré series. Math. Ann. 264(1), 9–20 (1983)
Holowinsky, R.: A sieve method for shifted convolution sums. Duke Math. J. 146(3), 401–448 (2009)
Ichino, A.: Trilinear forms and the central values of triple product L-functions. Duke Math. J. 145(2), 281–307 (2008)
Ingham, A.E.: Some asymptotic formulae in the theory of numbers. J. Lond. Math. Soc. 1(3), 202–208 (1927)
Ivić, A., Motohashi, Y.: The mean square of the error term for the fourth power moment of the zeta-function. Proc. Lond. Math. Soc. 69(2), 309–329 (1994)
Ivić, A., Motohashi, Y.: On some estimates involving the binary additive divisor problem. Q. J. Math. Oxford Second Ser. 46(184), 471–483 (1995)
Iwaniec, H.: Topics in Classical Automorphic Forms, Graduate Studies in Mathematics, vol. 17. American Mathematical Society, Providence (1997)
Iwaniec, H.: Spectral Methods of Automorphic Forms. Studies in Mathematics, vol. 53, 2nd edn. American Mathematical Society, Providence (2002)
Jutila, M.: The additive divisor problem and its analogs for Fourier coefficients of cusp forms. I. Math. Z. 223(3), 435–461 (1996)
Jutila, M.: On spectral large sieve inequalities. Funct. Approx. Comment. Math. 28, 7–18 (2000)
Kim, H., Sarnak, P.: Refined estimates towards the Ramanujan and Selberg conjectures, appendix to H H Kim, Functoriality for the exterior square of GL 4 and the symmetric fourth of GL 2, With appendix 1 by Dinakar Ramakrishnan and appendix 2 by Kim and Peter Sarnak. J. Am. Math. Soc. 16(1), 139–183 (2003)
Kuznetsov, N.V.: Petersson’s conjecture for cusp forms of weight zero and Linnik’s conjecture Sums of Kloosterman sums. Mat. Sb. (N.S.) 111(153), 334–383 (1980)
Lau, Y.-K., Liu, J., Ye, Y.: A new bound \(k^{2/3+\varepsilon }\) for Rankin–Selberg \(L\)-functions for Hecke congruence subgroups. Int. Math. Res .Papers 2006(35090), 35078–35090 (2006)
Liu, S.-C., Masri, R.: Nonvanishing of Rankin–Selberg L-functions for Hilbert modular forms. Ramanujan J. 34(2), 227–236 (2014)
Luo, W., Sarnak, P.: Quantum variance for Hecke eigenforms. Ann. Sci. École Norm. Sup. (4) 37(5), 769–799 (2004)
Montgomery, H.L., Vaughan, R.C.: Multiplicative Number Theory. I. Classical Theory, Cambridge Studies in Advanced Mathematics, vol. 97. Cambridge University Press, Cambridge (2007)
Motohashi, Y.: The binary additive divisor problem. Ann. Sci. École Norm. Sup. (4) 27(5), 529–572 (1994)
Motohashi, Y.: Spectral Theory of the Riemann Zeta-Function, Cambridge Tracts in Mathematics, vol. 127. Cambridge University Press, Cambridge (1997)
Nordentoft, A.C.: Central values of additive twists of modular L-functions. 1–40 (2018) arXiv:1812.08378
Nordentoft, A.C., Petridis, Y.N., Risager, M.S.: Small scale equidistribution of Hecke eigenforms at infinity (2020) arXiv:2011.05810 [math]
Phillips, R., Rudnick, Z.: The circle problem in the hyperbolic plane. J. Funct. Anal. 121(1), 78–116 (1994)
Rankin, R.A.: Contributions to the theory of Ramanujan’s function \(\tau \)(n) and similar arithmetical functions. II. The order of the Fourier coefficients of integral modular forms. Proc. Camb. Philos. Soc. 35, 357–372 (1939)
Sarnak, P.: Estimates for Rankin–Selberg L-functions and quantum unique ergodicity. J. Funct. Anal. 184(2), 419–453 (2001)
Selberg, A.: Bemerkungen über eine Dirichletsche Reihe, die mit der Theorie der Modulformen nahe verbunden ist. Arch. Mat. Naturvidenskab 43, 47–50 (1940)
Selberg, A.: On the Estimation of Fourier Coefficients of Modular Forms, vol. VIII, pp. 1–15. Proc. Sympos. Pure Math. Amer. Math. Soc., Providence (1965)
Watson, T.C.: Rankin Triple Products and Quantum Chaos, Ph.D. thesis, Princeton University (2002).
Acknowledgements
We are grateful to the anonymous referee for his/her many insightful comments and suggestions.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The research of Asbjørn C. Nordentoft and Morten S. Risager was supported by the Grant DFF-7014-00060B from Independent Research Fund Denmark.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Nordentoft, A.C., Petridis, Y.N. & Risager, M.S. Bounds on shifted convolution sums for Hecke eigenforms. Res. number theory 8, 26 (2022). https://doi.org/10.1007/s40993-022-00320-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s40993-022-00320-1