7 Large value estimates
The theory of zero density estimates for the Riemann zeta function (and other \(L\)-functions) rests on the study of what will be called large value patterns in this blueprint.
A large value pattern is a tuple \((N, T, V, (a_n)_{n \in [N,2N]}, J, W)\), where \(N {\gt} 1\) and \(T, V {\gt} 0\) are real numbers, \(a_n\) is a \(1\)-bounded sequence on \([N,2N]\), \(J\) is an interval of length \(T\), and \(W\) is a \(1\)-separated subset of \(J\) such that
for all \(t \in W\).
A Zeta large value pattern is a large value pattern in which \(J = [T,2T]\) and \(a_n = 1_I(n)\) for some interval \(I \subset [N,2N]\).
The choice of interval \(J\) is not too important for a large value pattern, since one can translate \(J\) and \(W\) by any shift \(t_0\) if we also modulate the coefficients \(a_n\) by \(n^{it_0}\) to compensate. However, this modulation freedom is not available for zeta large value patterns, as it destroys the form \(a_n = 1_I(n)\) of the coefficients. The cardinality \(|W|\) of \(W\) is traditionally called \(R\) in the literature.
It is common in the literature to relax the \(1\)-boundedness hypothesis on \(a_n\) slightly, for instance to \(a_n \ll T^{o(1)}\), but this does not significantly affect the analysis here. Similarly, the \(1\)-separation hypothesis is sometimes strengthened slightly to a \(\lambda \)-separation hypothesis for some \(\lambda = T^{o(1)}\), but again this does not make much difference. For some estimates, the uniform bound on \(a_n\) can be relaxed to an \(\ell ^2\) hypothesis \(\sum _{n \in [N,2N]} |a_n|^2 \ll N\) (and this second moment is traditionally called \(G\) in the literature), but we will not study such relaxations systematically here, as they are less relevant for the theory of zero density estimates.
Let \(1/2 \leq \sigma \leq 1\) and \(\tau \geq 0\) be fixed. We define \(\mathrm{LV}(\sigma ,\tau )\) to be the least fixed quantity for which the following claim is true: whenever \((N,T,V,(a_n)_{n \in [N,2N]},J,W)\) is a large value pattern with \(N{\gt}1\) unbounded, \(T = N^{\tau +o(1)}\), and \(V = N^{\sigma +o(1)}\), then
Implemented at large_values.py as:Large_Value_Estimate
One can check that the set of possible candidates for \(\mathrm{LV}(\sigma ,\tau )\) is closed (by underspill), non-empty, and bounded from below, so \(\mathrm{LV}(\sigma ,\tau )\) is well-defined as a real number. As usual, we have an equivalent non-asymptotic definition:
Let \(1/2 \leq \sigma \leq 1\), \(\tau \geq 0\), and \(\rho \geq 0\) be fixed. Then the following are equivalent:
\(\mathrm{LV}(\sigma ,\tau ) \leq \rho \).
For every (fixed) \(\varepsilon {\gt}0\) there exists \(C, \delta {\gt}0\) such that if \((N,T,V,(a_n)_{n \in [N,2N]},J,W)\) is a large value pattern with \(N \geq C\) and \(N^{\tau -\delta } \leq T \leq N^{\tau +\delta }\), \(N^{\sigma -\delta } \leq V \leq N^{\sigma +\delta }\), then one has
\[ |W| \leq C N^{\rho +\varepsilon }. \]
The proof of Lemma 7.3 is similar to that of Lemma 4.3, and is left to the reader.
(Monotonicity in \(\sigma \)) For any \(\tau \geq 0\), \(\sigma \mapsto \mathrm{LV}(\sigma ,\tau )\) is upper semicontinuous and monotone non-increasing.
(Huxley subdivision) For any \(1/2 \leq \sigma \leq 1\) and \(\tau ' \geq \tau \) one has
\[ \mathrm{LV}(\sigma ,\tau ) \leq \mathrm{LV}(\sigma ,\tau ') \leq \mathrm{LV}(\sigma ,\tau ) + \tau '-\tau . \]In particular, \(\tau \mapsto \mathrm{LV}(\sigma ,\tau )\) is Lipschitz continuous.
(\(\tau =0\) endpoint) One has \(\mathrm{LV}(\sigma ,0)=0\) for all \(1/2 \leq \sigma \leq 1\), and hence by (ii) \(0 \leq \mathrm{LV}(\sigma ,\tau ) \leq \tau \) for all \(1/2 \leq \sigma \leq 1\) and \(\tau \geq 0\).
TODO: implement Huxley subdivision as a way to transform a large values estimate into a better estimate
All claims are clear except perhaps for the upper bound
but this follows because any interval of length \(N^{\tau '+o(1)}\) may be subdivided into \(N^{\tau '-\tau +o(1)}\) intervals of length \(N^{\tau +o(1)}\), so on applying Definition 7.2 to each subinterval and summing (using Lemma 2.1 to ensure uniformity), one obtains the claim.
For any \(1/2 {\lt} \sigma \leq 1\) and \(\tau \geq 0\), one has \(\mathrm{LV}(\sigma , \tau ) \geq \min (2 - 2 \sigma ,\tau )\), while for \(\sigma =1/2\) one has \(\mathrm{LV}(\sigma ,\tau ) = \tau \).
In view of Lemma 7.4(ii), it suffices to show that \(\mathrm{LV}(\sigma , 2-2\sigma ) \geq 2-2\sigma \). By definition, it suffices to find a large value pattern \((N,T,V,(a_n)_{n \in [N,2N]},J,W)\) is a large value pattern with \(N\) unbounded, \(V=N^{\sigma +o(1)}\), \(T = N^{2-2\sigma +o(1)}\), and \(|W| \gg N^{2-2\sigma -o(1)}\).
In the endpoint case \(\sigma = 1\) one can achieve this by setting \(a_n=1\) for all \(n\) and taking \(W=\{ 0\} \), so now we assume that \(1/2 {\lt} \sigma {\lt}1\).
We use the probabilistic method. We divide \([N,2N]\) into \(\asymp N^{2-2\sigma }\) intervals \(I\) of length \(\asymp N^{2\sigma -1}\). On each interval \(I\), we choose \(a_n\) to equal some randomly chosen sign \(\epsilon _I \in \{ -1,+1\} \), with the \(\epsilon _I\) chosen independently in \(I\). If \(t = o(N^{2-2\sigma })\), then \(\sum _{n \in I} a_n n^{-it}\) is equal to \(\epsilon _I\) times a deterministic quantity \(c_{t,I}\) of magnitude \(\asymp N^{2\sigma -1}\) (the point being that the phase \(t \log n\) is close to constant in this range). By the Chernoff bound, we thus see that for any such \(t\), \(\sum _{n \in [N,2N]} a_n n^{it}\) will have size \(\gg N^{(2\sigma -1) + (2-2\sigma )/2} = N^\sigma \) with probability \(\gg 1\). By linearity of expectation, we thus see that with positive probability, a \(\gg 1\) fraction of integers \(t\) with \(t = o(N^{2-2\sigma })\) will have this property, giving the claim.
Finally, let \(\sigma =1/2\). In this case we just take each \(a_n\) to be a random sign, then by the Chernoff bound one has for each \(t\) that \(|\sum _{n \in [N,2N]} a_n n^{it}| \asymp N^{1/2}\) with positive probability, which by linearity of expectation as before gives the lower bound \(\mathrm{LV}(\sigma ,\tau ) \geq \tau \), while the upper bound is trivial from Lemma 7.4(iii).
We conjecturally have a complete description of the function \(\mathrm{LV}\):
One has
for all fixed \(1/2 {\lt} \sigma \leq 1\) and \(\tau \geq 0\). Equivalently (by Lemma 7.4(ii), (iii) and Lemma 7.5), one has \(\mathrm{LV}(\sigma , \tau ) = \min (2 - 2 \sigma ,\tau )\) for all \(1/2 {\lt} \sigma \leq 1\) and \(\tau \geq 0\).
Implemented at large_values.py as:montgomery_conjecture
We refer to [ 19 ] for further discussion of this conjecture, including some counterexamples to strong versions of the conjecture in which certain epsilon losses are omitted. In view of this conjecture, we do not expect any further lower bounds on \(\mathrm{LV}(\sigma ,\tau )\) to be proven, and the literature is instead focused on upper bounds.
The following application of subdivision is useful:
If \(\sigma \) is fixed, and the Montgomery conjecture holds for all fixed \(\tau {\lt} \tau _0\), then
for all fixed \(\tau \geq 0\).
Clear from Lemma 7.4(ii).
The following basic property of \(\mathrm{LV}(\sigma ,\tau )\) is extremely useful in applications:
For any \(1/2 \leq \sigma \leq 1\), \(\tau \geq 0\), and natural number \(k\), one has
Implemented at large_values.py as:raise_to_power_hypothesis()
Let \((N,T,V,(a_n)_{n \in [N,2N]},J,W)\) be a large value pattern with \(T = N^{k\tau +o(1)}\) and \(V = N^{\sigma +o(1)}\), Raising 1 the \(k^{\mathrm{th}}\) power, we conclude that
for all \(t \in W\), where \(b_n\) is the Dirichlet convolution of \(k\) copies of \(a_n\), and thus is bounded by \(N^{o(1)}\) thanks to divisor bounds. Subdividing \([N^k, 2^k N^k]\) into \(k\) intervals of the form \([N',2N']\) for \(N' \asymp N^k\) and applying Definition 7.2 (with \(N, T, V\) replaced by \(N', T, V^k\)) we conclude that
and the claim then follows.
7.1 Known upper bounds on \(\mathrm{LV}(\sigma ,\tau )\)
Similarly to upper bounds on \(\beta (\alpha )\), upper bounds on \(\mathrm{LV}(\sigma ,\tau )\) in the literature (also known as large value theorems) tend to be piecewise linear functions of \(\sigma \) and \(\tau \). Such bounds often tend to be convex initially (i.e., the maximum of several linear functions), but when one combines multiple large value theorems together, the bound is usually neither convex nor concave, though it often remains piecewise linear, and continuous in \(\tau \) (though jump discontinuities in \(\sigma \) are possible).
Listed below are some examples of such bounds.
For any fixed \(1/2 \leq \sigma \leq 1\) and \(\tau \geq 0\) one has
In particular, the Montgomery conjecture 2 holds for \(\tau \leq 1\).
Implemented at large_values.py as:large_value_estimate_L2
Let \((N,T,V,(a_n)_{n \in [N,2N]},J,W)\) be a large value pattern with \(T = N^{\tau + o(1)}\), \(V = N^{\sigma + o(1)}\). Applying [ 149 , Theorem 9.4 ] (with \(N\), \(T\) replaced with \(2N\), \(2T\) respectively and taking \(a_n = 0\) for \(n {\lt} N\)) one has
from which the result follows.
If \(1/2 \leq \sigma \leq 1\) and \(\tau \geq 0\) is such that
(this condition is vacuous for \(\tau {\lt} 1\)) then the Montgomery conjecture 2 holds for this choice of parameters.
For a stronger version of this inequality, see Lemma 8.12.
Set \(\rho := \mathrm{LV}(\sigma ,\tau )\); we may assume without loss of generality that \(\rho \geq 0\). Then by Definition 7.2, we can find a large value pattern \((N,T,V,(a_n)_{n \in [N,2N]},J,W)\) with \(N {\gt} 1\) unbounded, \(T = N^{\tau + o(1)}\), \(V = N^{\sigma + o(1)}\), and \(|W| = N^{\rho +o(1)}\). From 1 we have
hence for some \(1\)-bounded coefficients \(c_t\)
We apply the Halász argument. Interchanging the summations and applying Cauchy–Schwarz, we conclude that
hence on squaring and using the triangle inequality
In the case \(|t - t'| \leq N^{1-\varepsilon }\) for any fixed \(\varepsilon {\gt}0\), one can use Lemma 4.4 to obtain the bound
The total contribution of this case can then be bounded by \(N^{1+o(1)} R = N^{1+\rho +o(1)}\), thanks to the \(1\)-separation. In the remaining cases \(|t - t'| \geq N^{1-o(1)}\), we use Definition 4.2 to see that
and thus
By hypothesis, the second term on the right-hand side is asymptotically smaller than the left-hand side, and so we obtain \(\rho \leq 2-2\sigma \) as required.
If \((k,\ell )\) is an exponent pair, and \(1/2 \leq \sigma \leq 1\), and \(\tau \geq 0\) are fixed, then
In particular, the Montgomery conjecture holds for \(\tau \leq \frac{2\sigma +k-\ell -1}{k}\).
One can also obtain a similar implication starting from a bound on \(\mu \): see Lemma 8.13.
[ 122 , Equation (2.9) ] Let \(1/2 \leq \sigma \leq 1\) and \(\tau \geq 0\) be fixed. Then one has
In particular, one has the Montgomery conjecture for \(\tau \leq 4 \sigma - 2\).
Recorded in literature.py as:add_huxley_large_values_estimate()
Let \(1/2 \leq \sigma \leq 1\) and \(\tau \geq 0\) be fixed. If \(\mathrm{LV}(\sigma , \tau ) \le \rho \) then
Follows from [ 107 , Lemma 1 ] .
Let \(1/2 \leq \sigma \leq 1\) and \(\tau \geq 0\) be fixed. One has
In particular, the Montgomery conjecture holds for \(\tau \leq 11 \sigma - 8\).
Recorded in literature.py as:add_heath_brown_large_values_estimate()
By Lemma 7.7 it suffices to show that \(\mathrm{LV}(\sigma ,\tau ) \leq 2-2\sigma \) for \(\tau \leq 11\sigma -8\). From the previous theorem, and setting \(\rho = \mathrm{LV}(\sigma ,\tau )\), we have either
or
The latter bound can be rearranged as
and thus
and the claim follows. (See also the arguments in the first paragraph of [ 107 , p. 226 ] .)
If \(3/4 {\lt} \sigma \leq 1\) and \(\tau \geq 0\) are fixed, then
for any positive integer \(k\).
Let \((N,T,V,(a_n)_{n \in [N,2N]},J,W)\) be a large value pattern with \(N \geq 1\) be unbounded, \(T = N^{\tau + o(1)}\), \(V = N^{\sigma + o(1)}\), and \(|W| = N^{\rho +o(1)}\). By [ 106 , Lemma 6 ] we have
and thus
Since \(\sigma {\gt} 3/4\), we can delete the second term \(2\rho +1/2\) on the right-hand side. Solving for \(\rho \), we conclude that
and taking suprema in \(\rho \), we obtain the claim.
For any integer \(k \geq 1\), one has
Thus for instance with \(k=2\) we have
and with \(k=3\) we have
In particular, the Montgomery conjecture holds for
Recorded in literature.py as:add_jutila_large_values_estimate(Constants.LARGE_VALUES_TRUNCATION)
See [ 160 , (1.4) ] (setting \(V = N^{\sigma +o(1)}\), \(T = N^{\tau +o(1)}\), and \(G \leq N\)). We remark that this form is an optimized form of the inequality after (3.2) in Jutila’s paper, which in our notation would read that
whenever \(\mathrm{LV}(\sigma ,\tau ) \leq \rho \). The optimization follows from Lemma 7.7 and routine calculations.
Some additional large values theorems are established in Chapter 10.