Shrinking archimedean families: second moment for \({\mathop{\mathrm{GL}}}_2\)

TH, PN

Abstract

We attempt, unsuccessfully, to estimate certain second moments for \({\mathop{\mathrm{GL}}}_2\) involving conductor-truncated families.

§1. Overview

We consider \(\pi\) on \({\mathop{\mathrm{PGL}}}_2(\mathbb{Z}) \backslash {\mathop{\mathrm{PGL}}}_2(\mathbb{R})\) and try to estimate \[\sum_{C(\pi) \leq Q} \left\lvert L (\pi, \tfrac{1}{2} + i T ) \right\rvert^2.\] Here \(Q\) and \(T\) are asymptotic parameters. We have in mind the range \(Q \ll T\). In this range (or indeed, for \(Q \lll T^2\)), the analytic conductors for the individual \(L\)-functions are \(\asymp T^2\), so the convexity bound for the the squared \(L\)-function is \(\ll T\). It is straightforward to obtain an asymptotic formula for the above moment in the range where \(Q \ggg T\). We would like to obtain an essentially sharp upper bound for some \(Q \lll T\), ideally \(Q \ll T^{1-\delta}\). This seems hard.

Note that the range \(Q \asymp T\) is critical: a sharp bound for the moment in this range recovers the convexity bound for the individual \(L\)-values, while a sharp bound in any shorter range would give a subconvex bound.

§2. Test functions

Let’s set things up. We take a test function \(f_0\) to be a normalized smoothened characteristic function of \(K_0(Q)\), the archimedean variant of the standard congruence subgroup, like in (Jana and Nelson 2019): \[K_0(Q) = \left\{ \begin{pmatrix} a & b \\ c & d \\ \end{pmatrix} : a = 1 + o(1), \quad b = o(1), \quad c \lll 1/Q, \quad d = 1 + o(1) \right\}.\]

This should typically pick off something like an “analytic newvector” \(W_0 \in \pi\) for \(\pi\) with \(C(\pi) \leq Q\). In the Kirillov model, \(W_0\) could be taken to look like a smooth bump supported near \(1\): \[W_0(y) \approx 1_{y \asymp 1}^{\text{smooth}}.\] We then defined \(f\) to be the conjugate of \(f_0\) by \(n(T)\). On the spectral side, the contribution from \(\pi\) will be \[\left\lvert L(\pi, \tfrac{1}{2} + i T) \right\rvert^2 \sum _{W_0 \in \mathcal{B}(\pi)} \left\lvert \int _{y \in \mathbb{R} ^\times } \pi(f) W_0(y) |y|^{i T} \, d^\times y \right\rvert^2.\] Now consider the contribution from an “analytic newvector” \(W_0\) as above. The local weight will be, with \[W := n(T) W_0, \quad W(y) = e(T y) W_0(y),\] \[\left\lvert \int _{y \in \mathbb{R}^\times } W (y) |y|^{i T} \, d^\times y \right\rvert^2 \asymp T^{-1}.\] So far, so good.

§3. Geometric approximate functional equation

The problem is that we’re looking at the global period integral: for \(\varphi \in \pi\), \[\int _{y \in \mathbb{R}^\times / \mathbb{Z} ^\times } \varphi(a(y)) |y|^{i T} \, d^\times y, \quad a(y) := \begin{pmatrix} y & 0 \\ 0 & 1 \end{pmatrix}.\] We really want to replace this with an integral over a compact subset of \(\mathbb{R}^\times / \mathbb{Z}^\times\), so that we can later apply Cauchy—Schwarz productively. We argue like in (Michel and Venkatesh 2010, sec. 5.1.4) (see also (Nelson 2021, sec. 5.3)). The idea is that if we smoothen this integral out, then we get quite bounds away from some critical dyadic range, and then we focus on that range.

Let’s get started by fixing \(h \in C_c^\infty(\mathbb{R}^\times_+)\). We Mellin expand \(h\): \[h(t) = \int _{(\sigma)} H(s) t^s \, \frac{d s }{ 2 \pi i}.\] We assume \(h\) normalized to have integral one, so that \(H(0) = 1\).

For each positive parameter \(Y \in \mathbb{R}^\times_+\), consider \[I(Y) := \int _{y \in \mathbb{R}^\times / \mathbb{Z} ^\times } h \left( \frac{\lvert y \rvert}{Y} \right) \varphi(a(y)) |y|^{i T} \, d^\times y.\] Then we aim to bound \(I(Y)\) using the convexity bound for \(L(\pi,s)\). We have \[\label{eqn:cool-integral-rep-of-I-of-Y}\tag{1} I (Y) = \int _{(\sigma)} Y^{-s} \tilde{I}(s) \, \frac{d s}{2 \pi i},\] where \[\tilde{I}(s) :=H(s) Z(\varphi,\tfrac{1}{2} + s+ iT),\] \[Z(\varphi,\tfrac{1}{2} + s) := \int _{y \in \mathbb{R}^\times / \mathbb{Z}^\times } \varphi (a (y)) \lvert y \rvert ^s \, d^\times y.\] \(H(s)\) decays rapidly, so we can think of it informally as truncating to \(s = O(1)\).

Strategy: eventually we will bound \(\tilde{I}(0) = Z(\varphi, \tfrac{1}{2} + i T)\) by applying Cauchy’s theorem: \[\tilde{I}(0) = \oint \frac{\tilde{I}(s)}{s} \, \frac{d s }{2 \pi i},\] where, since \(\tilde{I}\) decays rapidly, we can take the contour to consist of a vertical line at \(\Re(s) = \varepsilon\) going up followed by a vertical line at \(\Re(s) = - \varepsilon\) going down, i.e., we consider the “box” \[\varepsilon- i \infty \rightarrow \varepsilon+ i \infty \rightarrow - \varepsilon+ i \infty \rightarrow - \varepsilon- i \infty \rightarrow \varepsilon- i \infty.\] Here we will have \(s \gg 1\), and also \(\tilde{I}(s)\) will decay rapidly, so the main point is to bound, for \(\Re(s) = \pm \varepsilon\), \[\tilde{I}(s) = \int _{Y \in \mathbb{R}^\times_+} Y^{-s} I(Y) \, \frac{d Y }{ Y},\] which, by the triangle inequality, satisfies \[\left\lvert \tilde{I}(s) \right\rvert \leq \int _{Y \in \mathbb{R}^\times_+} \max(Y,1/Y)^{\varepsilon} \lvert I(Y) \rvert \, \frac{d Y }{Y}.\]

Note: the bound that we seek for \(\tilde{I}(0)\) should be compared to the trivial bound following from convexity, which is \[\tilde{I}(0) \asymp T^{-1/2} L(\pi, \tfrac{1}{2} + iT) \prec 1.\]

We need to bound \(Z(\varphi,s)\). We do this via interpolation. In general, \[Z(\varphi,\tfrac{1}{2} + s) = L(\pi, \tfrac{1}{2} + s) Z(W, \tfrac{1}{2} + s),\] where \(W = W_\varphi\) is as constructed above and \[Z(W, \tfrac{1}{2} + s) = \int _{\mathbb{R}^\times } W_0(y) e(T y) \lvert y \rvert ^s \, d ^\times y.\] For \(\Re(s) \ll 1\), since \(W\) is a bump near \(1\), we have \[Z(W,\tfrac{1}{2} + s) \approx T^{-1/2 - \Im(s)} 1_{\Im(s) \asymp T}.\] So if we take \(\Re(s) = 1/2 + \varepsilon\), then we get a bound of \(\ll T^{-1/2}\). On the other hand, by the convexity bound, \[L(\pi, \tfrac{1}{2} + s) \prec 1 \text{ for } \Re(s) = 1/2 + \varepsilon.\] So this tells us that \[Z(\varphi, \tfrac{1}{2} + s) \prec T^{-1/2} \text{ for } \Re(s) = 1/2 + \varepsilon.\]

What does this tell us concretely? Look back at the integral representation \((1)\). If we shift to \(\sigma = 1/2 + \varepsilon\), then the function \(H(s)\) will truncate us to \(s \ll 1\), so we can bound the integral by something like its pointwise values at \(s \ll 1\), which will be \[\prec Y^{-1/2} T^{-1/2}.\] What this is saying is that if \(Y\) is a bit larger than \(T^{-1}\), then the “trivial bound” for \(I(Y)\) that we just sketched is stronger than \(1\). So it suggests that the main range to consider will be when \(Y \lessapprox T^{-1}\).

Now we should do the same thing but shifting in the opposite direction to find a complementary upper bound on the range of \(Y\) that we need to consider. Let’s shift to \(\Re(s) = -1/2 - \varepsilon\) for small \(\varepsilon> 0\). Then we have, for \(s \ll 1\), \[L(\pi, \tfrac{1}{2} + s + i T) \prec T,\] while we get the same bound \(Z(W, \tfrac{1}{2} + s + iT) \prec T^{-1/2}\) as before. Thus \[Z(\varphi, \tfrac{1}{2} + s) \prec T^{1/2} \text{ for } \Re(s) = -1/2 - \varepsilon.\] Now again shifting to \(\sigma = -1/2 - \varepsilon\) in \((1)\), we get \[I(Y) \prec Y ^{1/2} T ^{1/2}.\] This bound will be stronger than “\(\prec 1\)” if \(Y\) is a bit smaller than \(T^{-1}\).

Thus, the moral is that if we just want a subconvex bound for \(L(\pi,\tfrac{1}{2} + i T)\), then it suffices to nontrivially estimate \(I(Y)\) for \(Y \approx 1/T\), i.e., up to \(T^\varepsilon\) factors. Of course to actually recover the Weyl bound we need to consider a wider range of \(Y\) and make the analysis uniform in that. We would have had to do the same thing in the “classical” approach; the corresponding feature there is that the approximate functional equation has smaller dyadic ranges than the main one, i.e., we have \[L(\pi, \tfrac{1}{2} + i T) \approx \sum _{n \ll T} \frac{\lambda(n) }{n ^{1/2 + i T}},\] which we can’t altogether approximate by the contribution from \(n \asymp T\).

§4. Applying relative trace formula

So we should now, I think, study \(I(Y)\), for \(Y \approx 1/T\), via relative trace formula whatever stuff. That means we should write down the double integral (\(H = {\mathop{\mathrm{GL}}}_1 \hookrightarrow {\mathop{\mathrm{PGL}}}_2\)) \[\int _{ \substack{ x, y \in H : \\ x, y \asymp 1/T } } \lvert x/y \rvert ^{i T} \sum _{\gamma \in \Gamma } f (x ^{-1} \gamma y) \, d x \, d y.\] Here \(d x\) and \(d y\) denote Haar measures on \(H\), i.e., of the form \(d t / |t|\) with respect to Lebesgue measure, so that the integral over \(x\) and \(y\) is roughly a probability measure. This sum should correspond very roughly to \[\sum _{C(\pi) \leq Q} T^{-1} \left\lvert L(\pi, \tfrac{1}{2} + i T) \right\rvert^2,\] or at least the “main dyadic part” of those \(L\)-values. It may be useful to write \(x, y\) as multiples by \(a(1/T)\) over elements in \(H\) of size \(\asymp 1\), so that the main thing to consider becomes \[\int _{ \substack{ x, y \in H : \\ x, y \asymp 1 } } \lvert x/y \rvert ^{i T} \sum _{\gamma \in \Gamma } f (a(T) x ^{-1} \gamma y a(1/T) ) \, d x \, d y.\] We want to bound this by \(\ll Q/T\).

Remark 1. More precisely, here an expression like \[\int_{ \substack{ x \in H : \\ x \asymp 1 } } f(x) \, d x\] means \[\int_{x \in H \cong \mathbb{R}^\times } f(x) V(x) \, d x,\] where \(V\) lies in some fixed bounded subset of \(C_c^\infty(\mathbb{R}^\times)\). For example, we could take \(V\) to be a fixed element of that space, such as a smooth bump function supported on the interval \((1,2)\).

§5. Writing stuff out

We remember that \[f (g) = f _0 (n (- T) g n (T)).\] Thus \[f (a(T) x ^{-1} \gamma y a(1/T) ) = f_0 (n(-T) a(T) x ^{-1} \gamma y a(1/T) n(T)).\] We can do some conjugation: \[n(-T) a(T) x ^{-1} \gamma y a(1/T) n(T) = a(T) n(-1) x ^{-1} \gamma y n(1) a(1/T).\] \(f_0\) should detect when this lands in \(K_0(Q)\). \[K_0(Q) = G \cap \left( 1 + \begin{pmatrix} o(1) & o(1) \\ o(1/Q) & o(1) \end{pmatrix} \right),\] so \[a(1/T) K_0(Q) a(T) = K_0(Q) = G \cap \left( 1 + \begin{pmatrix} o(1) & o(1/T) \\ o(T/Q) & o(1) \end{pmatrix} \right).\] So the main condition to work with is now that \[n(-1) x ^{-1} \gamma y n (1) \in 1 + \begin{pmatrix} o(1) & o(1/T) \\ o(T/Q) & o(1) \end{pmatrix} =: J.\]

There’s the contribution from \(\gamma \in \Gamma_H \cong \{\pm 1\}\). For this, we’re basically looking at \[Q \int _{ \substack{ x \in H : \\ x \asymp 1 } } 1 _{n(-1) x n (1) \in J} \, d x.\] We have \[n (-1) x n (1) = \begin{pmatrix} x & x-1 \\ 0 & 1 \end{pmatrix}.\] This lies in \(J\) only if \(x = 1 + o(1/T)\), which happens with probability \(\lll T\), so we get the required bound \(Q/T\).

It remains to estimate the contribution of the off-diagonal: \[Q \sum _{\gamma \in \Gamma - \Gamma_H} \int _{ \substack{ x, y \in H : \\ x, y \asymp 1 } } \lvert x/y \rvert^{i T} 1 _{n (- 1 ) x ^{-1} \gamma y n (1) \in J} \, d x \, d y.\] We’ll see below that we’re in a range where it’s not possible to extract oscillation from the integrals over \(x\) and \(y\).

§6. Matrices

Thus \[f (a(T) x ^{-1} \gamma y a(1/T) ) = f_0 (n(-T) a(T) x ^{-1} \gamma y a(1/T) n(T)).\] We can do some conjugation: \[n(-T) a(T) x ^{-1} \gamma y a(1/T) n(T) = a(T) n(-1) x ^{-1} \gamma y n(1) a(1/T).\] \(f_0\) should detect when this lands in \(K_0(Q)\). \[K_0(Q) = G \cap \left( 1 + \begin{pmatrix} o(1) & o(1) \\ o(1/Q) & o(1) \end{pmatrix} \right),\] so \[a(1/T) K_0(Q) a(T) = G \cap \left( 1 + \begin{pmatrix} o(1) & o(1/T) \\ o(T/Q) & o(1) \end{pmatrix} \right).\] So the main condition to work with is now that \[n(-1) x ^{-1} \gamma y n (1) \in 1 + \begin{pmatrix} o(1) & o(1/T) \\ o(T/Q) & o(1) \end{pmatrix} =: J.\]

There’s the contribution from \(\gamma \in \Gamma_H \cong \{\pm 1\}\). For this, we’re basically looking at \[Q \int _{ \substack{ x \in H : \\ x \asymp 1 } } 1 _{n(-1) x n (1) \in J} \, d x.\] We have \[n (-1) x n (1) = \begin{pmatrix} x & x-1 \\ 0 & 1 \end{pmatrix}.\] This lies in \(J\) only if \(x = 1 + o(1/T)\), which happens with probability \(\lll T\), so we get the required bound \(Q/T\).

It remains to estimate the contribution of the off-diagonal: \[Q \sum _{\gamma \in \Gamma - \Gamma_H} \int _{ \substack{ x, y \in H : \\ x, y \asymp 1 } } \lvert x / y \rvert^{i T} 1 _{n (- 1 ) x ^{-1} \gamma y n (1) \in J} \, d x \, d y.\] We want to bound this by \(\ll Q/T\)? The convexity bound for \(\lvert L \rvert^2\) is \(\ll T\), so we need to bound the sum by \(\lll 1\) to improve upon convexity. So we really need to show \[\sum _{\gamma \in \Gamma - \Gamma_H} \int _{ \substack{ x, y \in H : \\ x, y \asymp 1 } } \lvert x / y \rvert^{i T} 1 _{n (- 1 ) x ^{-1} \gamma y n (1) \in J} \, d x \, d y \lll 1/Q\] but we might hope to be able to show (for certain ranges of \(Q\)) \[\sum _{\gamma \in \Gamma - \Gamma_H} \int _{ \substack{ x, y \in H : \\ x, y \asymp 1 } } \lvert x / y \rvert^{i T} 1 _{n (- 1 ) x ^{-1} \gamma y n (1) \in J} \, d x \, d y \ll 1/T.\] We recall that, writing \[\gamma = \begin{pmatrix} a & b \\ c & d \end{pmatrix},\] we have \[x ^{-1} \gamma y = \begin{pmatrix} a y/x & b/x \\ c y & d \end{pmatrix},\] hence \[n (-1) x ^{-1} \gamma y n (1) = \begin{pmatrix} a y/x - c y & b/x - d + a y / x - c y \\ c y & d + c y \end{pmatrix}.\] We arrive at the following conditions:

  1. \(c y = o(T/Q)\), or equivalently, \(c \lll T/Q\) (because \(y \asymp 1\)),

  2. \(ay / x - c y = 1 + o(1)\), which determines \(a\) up to \(o(1)\) if we know \((x,y,c)\),

  3. \(d + c y = 1 + o(1)\), which determines \(d\) up to \(o(1)\) if we know \((x,y,c)\),

  4. \(b /x - d + a y / x - c y = o(1/T)\), which should be satisfied about a proportion \(1/T\) of the time.

So that would lead to an overall bound of \(1 / Q\). We need to do a bit better than that. Seems tough!

Jana, Subhajit, and Paul D. Nelson. 2019. Analytic newvectors for \(\mathrm{GL}_n(\mathbb{R})\).” arXiv e-Prints, November, arXiv:1911.01880. https://arxiv.org/abs/1911.01880.
Michel, Philippe, and Akshay Venkatesh. 2010. “The Subconvexity Problem for \({\rm GL}_2\).” Publ. Math. Inst. Hautes Études Sci., no. 111: 171–271. https://doi.org/10.1007/s10240-010-0025-8.
Nelson, Paul D. 2021. Bounds for standard \(L\)-functions.” arXiv e-Prints, September, arXiv:2109.15230. https://arxiv.org/abs/2109.15230.