Aarhus Automorphic Forms Conference

Abstract

Partial notes on some of the lectures. These notes are incomplete, have not been proofread, and should be considered only a crude approximation to what happened in the lectures, filtered through my own misunderstandings and distractions. Any errors should be assumed to be due to the note-taker. Corrections welcome!

Peter Koymans, Hilbert 10 via additive combinatorics

Let’s start with a question of Hilbert from the influential 1900 conference:

Question 1. Does there exist an algorithm with

Let’s give a brief overview of the history of this question:

Definition 2. A subset \(S \subseteq \mathbb{Z}\) is called Diophantine if there exists an integer \(m > n\) and a polynomial \(f \in \mathbb{Z}[x_1, \dotsc, x_n]\) such that \[S = \left\{(a_1, \dotsc, a_n) \in \mathbb{Z}^n \mid \text{ there exists } a_{n + 1, \dotsc, a_m} \in \mathbb{Z}, \, f(a_1, \dotsc, a_m) = 0 \right\}.\] In other words, take the solution set to a polynomial equation and project it onto some coordinates.

A bit of history:

Theorem 3 (Mazur–Rubin). Let \(R\) be a finitely-generated ring over \(\mathbb{Z}\), with \(\lvert R \rvert = \infty\). Assume that for all number fields \(K\), and all elliptic curves \(E\) over \(K\), that the Tate–Shafarevich group \(\operatorname{Sha}(E/K)\) is finite. Then Hilbert’s 10th problem is undecidable over \(R\).

Theorem 4 (Koymans–Pagano, 2024). Same result, unconditionally.

Theorem 5 (Koymans–Pagano, 2024). Let \(K\) be a number field. Then \(\mathbb{Z}\) is Diophantine over \(\mathcal{O}_K\)

Let

Theorem 6 (Koymans–Pagano). Let \(K\) be a number field with \(\geq 32\) real embeddings. Then there exists \(E/K\) such that \(\operatorname{rank} E(K) = \operatorname{rank} E(K(i)) > 0\).

The proof that Theorem 6 implies Theorem 4 uses the following:

Lemma 7. Let \(K \subseteq L\) be number fields. Then:

  1. If \(D_1\), \(D_2\) are Diophantine over \(\mathcal{O}_K\), then so is \(D_1 \cap D_2\).

  2. If \(D \subseteq \mathcal{O}_K\) is Diophantine over \(\mathcal{O}_K\), and also \(\mathcal{O}_K\) is Diophantine over \(\mathcal{O}_L\), then \(D\) is Diophantine over \(\mathcal{O}_L\).

  3. If \(\mathbb{Z}\) is Diophantine over \(\mathcal{O}_L\), then \(\mathbb{Z}\) is Diophantine over \(\mathcal{O}_K\).

  4. If \(K\) is totally real, then \(\mathbb{Z}\) is Diophantine over \(\mathcal{O}_K\).

  5. If \(E\) satisfies \(\operatorname{rank} E(K) = \operatorname{rank} E(L) > 0\), then \(\mathcal{O}_K\) is Diophantine over \(\mathcal{O}_L\).

Assuming the above:

Proof that Theorem 6 implies Theorem 4. Let \(K\) be a number field. We want to show that \(\mathbb{Z}\) is Diophantine over \(\mathcal{O}_K\). Define \(M\) to be the Galois closure of \(K(i, \sqrt{2}, \dotsc, \sqrt{13})\). It suffices to show that \(\mathbb{Z}\) is Diophantine over \(\mathcal{O}_M\), by \((iii)\). Taking the decomposition group \(D \subseteq \mathop{\mathrm{Gal}}(M / \mathbb{Q})\) at an infinite place, we must have \(1 \notin M^D =: L\). Thus \(M = L(i)\). Hence by Theorem 6 and \((v)\), we know that \(\mathcal{O}_{M^D}\) is Diophantine over \(\mathcal{O}_M\). Hence by \((i)\), we see that \[\bigcap_{ \substack{ D \subseteq \mathop{\mathrm{Gal}}(M / \mathbb{Q}) \\ D \text{ decomp. at } \infty } } \mathcal{O}_D.\] is Diophantine over \(\mathcal{O}_M\). We may write the above as \(\mathcal{O}_F\), with \[F := \bigcap_{ \substack{ D \subseteq \mathop{\mathrm{Gal}}(M / \mathbb{Q}) \\ D \text{ decomp. at } \infty } } M^D.\] But \(F\) is totally real, so \(\mathbb{Z}\) is Diophantine over \(\mathcal{O}_F\) by \((iv)\), hence \(Z\) is Diophantine over \(\mathcal{O}_M\) by \((ii)\). ◻

Let’s now give the proof of Theorem 6.

We first discuss \(2\)-Selmer groups. Take the Galois cohomology of \(G_K := \mathop{\mathrm{Gal}}(\bar{K} / K)\). Look at \[0 \rightarrow E[2] \rightarrow E \xrightarrow{\bullet 2} E \rightarrow 0.\] This gives \[0 \rightarrow E(K)[2] \rightarrow E(K) \xrightarrow{\bullet 2} E(K) \xrightarrow{\delta} H^1(G_{K}, E[2]),\] so \[0 \rightarrow \frac{E(K)}{2 E(K)} \xrightarrow{\delta} H^1(G_K, E[2]).\] The first term is an \(\mathbb{F}_2\)-vector space of dimension \[\operatorname{rank} E(K) + \dim_{\mathbb{F}_2} E(K)[2].\] We can also do the same diagram for each place \(v\) of \(K\). You can check that this whole thing gives a commutative diagram \[\begin{CD} \small 0 @>>> E(K)[2] @>>> E(K) @>{\bullet 2}>> E(K) @>{\delta}>> H^1(G_K, E[2]) \\ @. @VVV @VVV @VV {\mathop{\mathrm{res}}}_vV @VV{\mathop{\mathrm{res}}}_vV \\ 0 @>>> E(K_v)[2] @>>> E(K_v) @>{\bullet 2}>> E(K_v) @>{\delta_v}>> H^1(G_{K_v}, E[2]) \end{CD}\]<++> \[0 \rightarrow E(K)[2] \rightarrow E(K) \xrightarrow{\bullet 2} E(K) \xrightarrow{\delta} H^1(G_K, E[2]) \xrightarrow{{\mathop{\mathrm{res}}}_v} H^1(G_{K_v}, E[2])\] and \[0 \rightarrow E(K_v)[2] \rightarrow E(K_v) \xrightarrow{\bullet 2} E(K_v) \xrightarrow{\delta_v} H^1(G_{K_v}, E[2]).\] Hence \({\mathop{\mathrm{Sel}}}^2(E / K) := \ker(H^1(G_K, E[2]) \xrightarrow{\prod {\mathop{\mathrm{res}}}_v} H^1(G_{K_v}, E[2]) / \mathop{\mathrm{image}}(\delta_v))\).

  1. If \(\mathop{\mathrm{image}}(\delta) \subseteq {\mathop{\mathrm{Sel}}}^2(E / K)\), then \(\mathop{\mathrm{rank}}E(K) + \dim_{\mathbb{F}_2} E(K)[2] \leq \dim_{\mathbb{F}_2} {\mathop{\mathrm{Sel}}}^2(E / K)\).

  2. \({\mathop{\mathrm{Sel}}}^2(E / K)\) is computable and finite-dimensional.

Let’s now discuss forcing positive rank. Take \(\tilde{E}\) with \(\tilde{E}(K)[2] \cong \mathbb{F}_2^2\), i.e., \(y^2 =(x - a_1)(x - a_2)(x - a_3)\) with \(a_1, a_2, a_3 \in K\) distinct. Twisting by \(t :=(n - a_1)(n - a_2)(n - a_3)\) gives a new elliptic curvge \(t y^2 =(x - a_1)(x - a_2)(x - a_3)\) with rational point \((x, y) =(0, 1)\). Twisting by \(t :=(n - a_1 d)(n - a_2 d)(n - a_3 d)\) gives a new curve with rational point \((x, y) =(\tfrac{n}{d}, \tfrac{1}{d^2})\).

Since \(i \notin K\), we have \(\operatorname{rank} \tilde{E}^t(K(i)) = \mathop{\mathrm{rank}}\tilde{E}^t(K)\) suffices to find \(t = d(n - a_1 d)(n - a_2 d)(n - a_3 d)\) with \(\mathop{\mathrm{rank}}\tilde{E}^{- t}(K) = 0\). It suffices to show that \(\dim_{\mathbb{F}_2} {\mathop{\mathrm{Sel}}}^2(E/K) = 2\), because this implies that \(\mathop{\mathrm{rank}}\tilde{E}^{- t}(K)[2] + 2 \leq 2\), whence \(\mathop{\mathrm{rank}}\tilde{E}^{- t}(K)[2] = 0\).

So we need to control \(2\)-Selmer groups. What’s the punchline? (We’ve talked about Hilbert 10, but not yet additive combinatorics.) Fact: we can compute \({\mathop{\mathrm{Sel}}}^2(E / K)\) from Legendre symbols of primes dividing \(t\).

Morten Risager, The error term in the hyperbolic circle problem

Let \(\Gamma \leq \mathrm{SL}_2(\mathbb{R})\) be a discrete, cofinite subgroup that contains \(- 1\). A good example is \(\mathrm{SL}_2(\mathbb{Z})\).

Let \(z, z ' \in \mathbb{H}\). We want to count how many \(\Gamma\)-translates of \(z\) lie in a hyperbolic circle based at \(z'\), with radius \(R \rightarrow \infty\): \[N(z, z', X) := \# \left\{ \gamma \in \Gamma \mid 4 u(\gamma z, z ') + 2 \leq X \right\}.\] Here \(u(z, z') = \frac{\lvert z - z ' \rvert^2}{4 \Im(z) \Im(z')}\). We note that \(4 u(\gamma z, z ') = 2 \cosh(d_H(\gamma z, z '))\), so one can formulate the counting problem in a few ways.

Selberg (1970’s) showed that \[N(z, z', X) = \frac{2 \pi}{\mathop{\mathrm{vol}}(\Gamma \backslash \mathbb{H})} X + (\text{contribution from small eigenvalues}) + \operatorname{O}(X^{2/3}).\] (Selberg never published this result, but he gave talks about it, and we have notes from these talks available at the IAS.)

Question 8. Can \(\operatorname{O}(X^{2/3})\) be improved? (Note that \(\tfrac{2}{3} = \tfrac{1}{2} + \tfrac{1}{6}\).)

We have the \(\Omega\)-result of Phillips–Rudnick (’94): \(N(z, z', X) - M(X) = \Omega(X^{\frac{1}{2} - \delta})\), for each \(\delta > 0\).

What do we do when we don’t know how to bound something? We try to do something easier: take an average of some sort. Look at the second radial moment.

Chamizo (’96): \[\left( \frac{1}{X} \int_X^{2 X} \left( N(z, z', X) - M(X) \right)^2 \, d x \right)^{1/2} = \operatorname{O}\bigl( X^{1/2} \log X\bigr).\] Cherubini (2016): \[\label{eq:cq6vse71xz}\tag{1} \operatorname{O}(X^{1/2} \log^{1/2} X).\]

Conjecture 9. \(N(z, z', X) - M(X) = \operatorname{O}(X^{\frac{1}{2} + \varepsilon})\).

Local averages. Biro (2018): take \(f\) of compact support, and \(z = z '\). Then \[\int_{\Gamma \backslash \mathbb{H}} f(z) \left( N(z, z, X) - M(X) \right) \, d \mu(z) = \operatorname{O}\bigl(X^{\frac{1}{2} + \frac{1}{8}}\bigr).\]

From now on, restrict to \(\Gamma = \mathrm{SL}_2(\mathbb{Z})\). In that case, the main term is simply \(6 X\). One can improve Phillips–Rudnick to \(\Omega\bigl(X^{\frac{1}{2}}(\log X)^{\frac{1}{4} - \delta}\bigr)\) , and Cherubini to \(\operatorname{O}(X^{\frac{1}{2} + \frac{1}{12} + \varepsilon})\).

We want now to restrict to an even more arithmetic situations, where \(z = z_d\) and \(z ' = z_{d '}\) are both Heegner points. We want \(d\) and \(d'\) to be different fundamental squarefree discriminants, with \(0 > d = b^2 - 4 a c \equiv 1 \pmod{4}\) and \[z_d = \frac{- b + \sqrt{\lvert d \rvert}}{2 a}.\]

Theorem 10. Second spacial moment: with \(\Gamma = \mathrm{SL}_2(\mathbb{Z})\), \[\Bigl( \int_{\Omega} \bigl( N(z, z, X) - M(X) \bigr)^2 \, d \mu(z) \Bigr)^{1/2} = \operatorname{O}(X^{\frac{1}{2} + \frac{1}{7}}).\]

Let’s return to the theorem:

Theorem 11 (Chatzakos, Cherubini, Lester, Risager, 2025). Let \(\Gamma = \mathrm{SL}_2(\mathbb{Z})\), and \(d\), \(d'\) are as above. Then we manage to prove \[N(z_d, z_{d'}, X) = b X + \operatorname{O}(\frac{X^{2/3}}{(\log X)^{1/6}}).\]

We can furthermore improve \((1)\) to \(\operatorname{O}\bigl(X^{1/2} \log^{3/8} X\bigr)\)

Corollary 12. Fix \(d\), \(d_0\) (not equal to \(-3\) (different squarefree fundmental discriminants). \[\eta_d = \left\{ p \mid a > 0, \, \text{discr}(p) = d, \, 0 <(\sigma - \operatorname{discr(p, p_0)})\right\} = \frac{6 h(d)}{ \sqrt{d d_0}} + \operatorname{O}(\frac{X^{2/3}}{(\log q k^{1/2})}).\]

To understand \(N(z, z', X)\), it is useful to understand bounds on \[S(x) := \sum_{T \leq t_j \leq 2 T} \sum_{T \leq t_j \leq 2 T} X^{t_j} \varphi_j(z) \overline{\varphi_j(z ')} \ll T^2.\] Here \(\lambda_j = \tfrac{1}{4} + j^2\) is an eigenvalue of \(\Delta_P\) with eigenfunction \(q_j(z)\).

We prove: \[\lvert S(X, T) \rvert \leq \sum_{T \leq t_j \leq 2 T} \left\lvert \varphi_{j}(z_d) \varphi_j(z_{d '}) \right\rvert \ll \frac{T^2}{(\log T)^{1/4}}.\]

Using Waldspurger: \[\lvert \varphi_j(z) \rvert \ll_d \sum_{\xi \in \widehat{\mathrm{Cl}(\mathbb{Q}(\sqrt{d}))}} \frac{L(\varphi_j \times f_\xi, \frac{1}{2})^{\frac{1}{2}}}{L(1, {\mathop{\mathrm{sym}}}^2 \varphi_j)}.\] Here \(f_s\) is the theta series related to the class group character \(\xi\).

Theorem 13. Let \(d\) and \(d'\) be different, \(\xi \in \widehat{\mathrm{Cl}(\mathbb{Q}(\sqrt{d}))}\) and \(\xi' \in \widehat{\mathrm{Cl}(\mathbb{Q}(\sqrt{d'}))}\). Then \[\sum_{T \leq t_j \leq 2 T} \frac{L(\varphi_j \times f_\xi, \frac{1}{2})^{1/2} L(\varphi_j, f_{\xi '})^{1/2}}{L({\mathop{\mathrm{sym}}}^2 \varphi_j, 1)} = \operatorname{O}\left( \frac{T^2}{(\log T)^{1/4}}\right)\]

In order to prove this, we needed \[\sum_{t_j} h(t_j) \frac{L(\tfrac{1}{2}, \varphi_j \times f_{\xi})}{L({\mathop{\mathrm{sym}}}^2 \varphi, 1)} \lambda_j(\ell) = C_h \frac{\lambda_{f_{\xi}}(\ell)}{\sqrt{\ell}} T^2 + \operatorname{O}(T^{1 + \varepsilon}).\] Here \(h\) localizes at \([T, 2 T]\).

Alan Zhao, A different perspective on Whittaker functions for \(\mathrm{GL}_n\)

Notation

Let \(\lambda\) denote a partition, say \((4 3 1)\), with accompanying Young diagram. We’ll also use \[(a;q) =(1 - a)(1 - a q) \dotsb(1 - a q^{n - 1}).\] We’ll also fix \(q < 1\).

To introduce the \(q\)-deformation Whittaker functions, we’ll start by defining:

MacDonald polynomials

These will lie in \(\mathbb{Q}(q, t)[x_1, \dotsc, x_n]\). They’ll depend upon a partition, whose length is the number of variables. For each \(\lambda\), there exists \(P_\lambda\) such that \(\langle P_\lambda, P_\mu \rangle = 0\) for \(\lambda \neq \mu\). We get \(P_\lambda(x, q, t)\). Specializations:

Example 14. The \(q\)-Whittaker polynomial associated to \(\lambda =(2 1)\) is \[(x_1^3 + x_2^3) +(1 + q + q t)(x_1^2 x_2 + x_1 x_2^2).\] If you just take \(q = 1\) naively, you get (...).)

\(q\)-Whittaker \(\Psi_{x_1, \dotsc, x_n}(\underline{\lambda}_n)\), where \(\underline{\lambda}_n\) is a Gelfand–Tsetlin pattern \(\lambda_{n, 1} \geq \lambda_{n, 2} \geq \dotsb \lambda_{n, n}\) with interlaced rows below that. This comes out to \[\Psi_{x_1, \dotsc, x_n}(\underline{\lambda}_n) = \sum_{\text{patterns}} \prod x_i^{\sum \lambda_{???} - \sum_{?}\lambda_{\dotsb}} \times(\text{quotient of }(q, q)_n).\]

Valentin Blomer, Hecke fields - Galois theory meets shifted convolution problems

Let \(f(z) = \sum_n a_n e(n z)\) be a cusp form, \(\mathbb{Q}(f) := \mathbb{Q}(\left\{ a_n \mid n \in \mathbb{N} \right\})\).

Let \(\chi\) be a Dirichlet character, \(\mathbb{Q}(\chi) := \mathbb{Q}(\left\{ \chi(n) \mid n \in \mathbb{N} \right\})\).

Shimura (1977): there exists \(\Omega_{f, \pm} \in \mathbb{C}^\times\) such that \[L^{\mathrm{alg}}(\tfrac{1}{2}, f \times \chi) := \frac{G(\bar{\chi}) L(\tfrac{1}{2}, f \times \chi)}{\Omega_{f, \mathop{\mathrm{sgn}}(\chi)}} \in \mathbb{Q}(f, \chi).\] Here we think of \(f\) as being fixed and \(\chi\) as varying. Moreover, we have the following reciprocity: for a Galois automorphism \(\sigma\), \[\sigma L^{\mathrm{alg}}(\tfrac{1}{2}, f \times \chi) := L(\tfrac{1}{2}, f^\sigma \times \chi^\sigma).\]

We recall that \[\mathbb{Z}_p^\times = \mu_{p - 1} \times(1 + p \mathbb{Z}_p) \cong \widehat{\mathbb{Z}^\times_p}.\] If we take a character \(\chi\) of prime power conductor \(p^{h}\), then \(p^{h - 1}\) divides the order of \(\chi\). We’re interested in the set of characters \(\chi\) such that the conductor is \(p^{h}\) and the order is \(p^{h - 1}\): \[\Xi_{p, h} := \left\{ \chi : \mathrm{conductor} = p^{h}, \mathrm{order} = p^{h - 1} \right\}.\] In the terminology of Ravi’s talk yesterday, these are “wild characters”. We’ll see later that this is a Galois orbit. The values of such a character \(\chi\) are \(p^{h - 1}\)th roots of unity, and most of these values are primitive. In particular, \[\operatorname{image}(\chi) = \mu_{p^{h - 1}}.\] More precisely, \(\chi(x)\) is a primitive \(p^{h - 1}\) root of unity if and only if \(x = x_0 + x_1 p + x_2 p^2 + \dotsb\), where \(p \nmid x_0 x_1\).

After this interlude on characters, we come to the thing we really want to discuss. The number \(L^{\mathrm{alg}}(\tfrac{1}{2}, f \times \chi)\) lies in the field \(\mathbb{Q}(f, \chi)\), and you might ask, is it a generator of the field? When is this the case?

The work of Luo–Ramakrishnan (1977) and later work of Sun (2019) established the following. Fix a prime \(p\) and a cusp form \(f\). We make the assumption that \(f\) has no inner twist, meaning that \(f\) is not equal to \(f \otimes \chi_D\) for a real character \(\chi_D\). (If the level is squarefree, then this can never happen.) Then they proved that \(\mathbb{Q}(f, \chi) = \mathbb{Q}(L^{\mathrm{alg}}(\tfrac{1}{2}, f \times \chi), \mu_p)\) holds for all \(\chi \in \Xi_{p, h}\) provided that \(h \geq h_0\). That’s quite remarkable - a central character is able to generate fields when you run through character.

We want to do the same with more complicated expressions involving this central value (see (Blomer et al. 2025)):

Theorem 15 (Blomer–Burungale–Michel–Min). Under the same assumptions, \(\mathbb{Q}(f, \chi) = \mathbb{Q}(\left\lvert L(\tfrac{1}{2}, f \times \chi) \right\rvert^2, \mu_p)\) for all \(\chi \in \Xi_{p, h}\) with \(h \geq h_0\).

We note that the containment “\(\supseteq\)” is obvious; the hard part is to show “\(\subseteq\)”.

As announced in the abstract, there is an analytic part and a Galois-theoretic part.

Galois theory

We have \(\mathbb{Q}(\chi) \subseteq \mathbb{Q}(\mu_p, \ast)\) and \(\mathbb{Q}(f) \subseteq \mathbb{Q}(\chi, \ast)\). The real reason we picked this topic is that we can now draw complicated field diagrams.

The top field is \(\mathbb{Q}(f, \chi)\), which contains \(\mathbb{Q}(\chi, \ast)\). This contains \(\mathbb{Q}(\chi)\) (branching down and to the left) and also \(\mathbb{Q}(\mu_p , \ast)\) (branching down and to the right), which both contain the common subfield \(\mathbb{Q}(\chi) \cap \mathbb{Q}(\mu_p, \ast)\), which in turn contains \(\mathbb{Q}\). There is a further extension, with dotted lines: \(\mathbb{Q}(f, \chi)\) contains \(\mathbb{Q}(f)\) which contains \(\mathbb{Q}\). Furthermore, \(\mathbb{Q}(f)\) is contained in its Galois extension \(E\), which is contained in (branching up and to the right) \(E(\chi)\), which contains (down and to the right) \(\mathbb{Q}(f, \ast)\), and then there’s the common subfield \(E \cap \mathbb{Q}(\chi, \ast)\) of \(E\) and \(\mathbb{Q}(f, \ast)\):

image

We set \(\ell_{f, \chi} := \lvert L^{\mathrm{alg}}(\tfrac{1}{2}, f \times \chi) \rvert^2\). By some diagram chasing, we have \[0 \neq {\mathop{\mathrm{trace}}}_{\mathbb{Q}(f, \chi) / \mathbb{Q}}(\chi(x) \ell_{f, \chi}) = {\mathop{\mathrm{trace}}}_{\mathbb{Q}(\mu_p, \ast) / \mathbb{Q}} \left( \ell_{f, \chi} {\mathop{\mathrm{trace}}}_{\mathbb{Q}(f, \chi) / \mathbb{Q}(\mu_p, \ast)} \chi(x) \right).\] We deduce that, unless the extension is trivial, we obtain \[\deg {\mathop{\mathrm{trace}}}_{\mathbb{Q}(\chi, \ast) / \mathbb{Q}(\mu_p, \ast)} \chi(x) = 0.\] Let \(\sigma\) be a Galois automorphism of \(E(\chi) / \mathbb{Q}(f, \ast)\). Then \[\ell_{f, \chi} = \sigma(\ell_{f, \chi}) = \ell_{f^2 \times \chi},\] which implies that \(f = f^\sigma\), so that the extension is trivial. We then consider \[{\mathop{\mathrm{trace}}}_{\mathbb{Q}(f_1, f_2, \chi) / \mathbb{Q}(f_1, f_2)}(\chi(x) \ell_{f_1, \chi}).\] The upshot is that we need to evaluate \[\label{eq:cq6v8d8hcu}\tag{2} \sum_{\sigma \in \mathop{\mathrm{Gal}}(F(\chi)) / F_0} L(\tfrac{1}{2}, f \times \chi^\sigma) \overline{F(\tfrac{1}{2}, g \times \chi^\sigma)} \chi^\sigma(x).\] We’re not summing over all characters, but only over a Galois orbit, so there is only some orthogonality: \[\mathbb{E}_{\sigma} \chi^\sigma(n) := \delta_{n^{p - 1} \equiv 1 \pmod{p^{h}}}.\]

Shifted convolution sums

The standard way to evaluate such a sum of \(L\)-values is via the approximate functional equation. This leads morally to something like \[\label{eq:cq6v8a8450}\tag{3} \sum_{ \substack{ m n \leq q^2 \\ \alpha m \equiv x n \pmod{q} } } \lambda_f(m) \lambda_f(n),\] where \(x\) is fixed, \(q = p^{h}\) and \(\alpha^{p - 1} \equiv 1 \pmod{q}\). The subtlety here is that \(\alpha\) is not fixed – there are only so many such roots of unity, but they’re spread out all over the place. Most results concerning the shifted convolution problem apply when the coefficients playing the role of \(\alpha\) and \(x\) in \((3)\) are essentially fixed.

Let’s start with the case \(\alpha = 1\). Take \(n \asymp N\), \(m \asymp M\).

  1. If \(N \leq q^{3/2 - \varepsilon}\), then one can apply spectral methods: \(\delta\)-symbol, Voronoi summation, large sieve. Indeed, this works as long as the longer variable is no more than three times as long as the shorter variable.

  2. If \(N \geq q^{3/2 + \varepsilon}\), then we apply Voronoi in \(n\), which leads to \(\frac{N}{q^2} \sum_{m \asymp M} \sum_{n \asymp q^2 / N} \lambda(m) \lambda(n) S(m, n, q)\), which we can treat using the Weil bound for Kloosterman sums.

  3. There remains the intermediary range. One can hope to treat this by upgrading the treatment of the \(N \geq q^{3/2 + \varepsilon}\) range, using that \(q = p^{h}\) is factorable, via Weyl differencing.

For \(\alpha \neq 1\) large, the treatment of case \((i)\) breaks down completely. Without a better idea how to proceed, we take absolute values: \[\sum_{(m, n) \in \Lambda}\lvert \lambda_f(n) \lambda_f(m) \rvert, \quad \Lambda := \left\{ \alpha m \equiv x n \pmod{q} \right\}.\] We observe that \(\Delta\) is a “roundish” lattice, and we’re summing nonnegative multiplicative functions over it. Sieve theory tells us that we understand this, and the answer doesn’t depend upon the shape of the lattice. Recall that (by approximations to Sato–Tate) the magnitudes \(\lvert \lambda_f(n) \rvert\) are a bit smaller than \(1\) on average. This saves \((\log q)^{- \delta}\) for some \(\delta \approx 0.1\).

This suffices for an asymptotic evaluation of the moment \((2)\). (We were worried that we already had to do a dyadic subdivision, so saving the fractional log would be swamped by the logarithmically many scales, but it turns out that the main term is of size \(q \log q\), so it really suffices.)

So, what is the upshot, or the moral to be learned? There’s somehow an interesting passage from generation of Hecke fields to moments – by a bit of Galois theory, we can relate such a generation question to moments of \(L\)-functions, at which point we can apply analytic theory. It results in situations where we’re summing over a Galois orbit, which changes the shifted convolution structure drastically.

Zvi Shem-Tov, Equidistribution of lifts on hyperbolic 4-manifolds

(Joint with de Faveri)

Take \(M = \Gamma \backslash \mathbb{H}^n\), where \(\Gamma = \mathbb{G}(\mathbb{Z})\) and \(G = \mathbb{G}(\mathbb{R}) = \mathop{\mathrm{SO}}(1, n)\).

Take a sequence \(\phi_j\) of joint eigenfunctions of the hyperbolic Laplacian \(\Delta\) and the family of Hecke operators \(\mathcal{H}\). So that this sequence is nontrivial, we require that \(\lvert \lambda_j \rvert \rightarrow \infty\). We also normalize so that \(\lVert \phi_j \rVert_2 = 1\). We define the probability measures \[\mu_j = \lvert \phi_j \rvert^2 \, d \mathop{\mathrm{vol}}.\]

Question 16. What are the weak-\(\ast\) limits of such \(\mu_j\)?

Conjecture 17 (Rudnick–Sarnak, “QUE”, 1995). The only such limit is the uniform measure \(\mu = d \mathop{\mathrm{vol}}\).

What is known so far? (We focus on this specific setup, ignoring many others.)

In the following work in progress, we can prove QUE in dimension \(4\) for a concrete, nontrivial example of such a sequence.

Theorem 18 (de Faveri–Shem-Tov). QUE holds for Pitale lifts. In fact, all we need to know is that the Hecke eigenvalues are large. The result applies, for instance, to \(\Gamma = \operatorname{SU}_2(\mathbb{Z})\).

Step 1 is the microlocal lift. Namely, we can lift the measure \(\mu\) to a measure \(\tilde{\mu}\) on the phase space \(\Gamma \backslash G\) in a way such that this lift satisfies two conditions:

We use a very general construction of Silberman–Venkatesh (although such constructions have a long history). That’s the first step – once we have this microlocal lift, we can forget the original quotients and just speak of measures on these homogeneous spaces.

Step 2 is ergodic theory or measure-rigidity. Basically, what we do is to try to classify these \(A\)-invariant measures, using entropy/recurrence conditions and results of Einsiedler–Lindenstrauss. What we get out is that \(\tilde{\mu}\) is a convex combination of homogeneous measures (i.e., a measure on the orbit of a reasonable reasonable). What we need to show is then that, for \(\pi : G \rightarrow \Gamma \backslash G\) the natural projection, we have \(\tilde{\mu}(\pi(L)) = 0\) for certain submanifolds \(L \subset G\). The most general result we know how to prove in this context is the following

Theorem 19 (Shem-Tov–Silberman). Assume that for any subspace \(U \subset L\) with irreducible Zariski closure \(\bar{U}^Z\), we have that \[S := \left\{ s \in \mathbb{G}(\bar{\mathbb{Q}} \cap \mathbb{R}) : s \bar{U}^Z = \bar{U}^Z \right\}\] is \(\tfrac{1}{2}\)-weakly small (to be defined below). Then \(\tilde{\mu}(\pi(L)) = 0\).

Weak smallness (defined by Marshall in the context of the sup norm problem). Let \(H < G\). Fix maximal split tori \(T_H < T_G\). For \(\eta \in X_\ast(T_G)\), define \[\lVert \eta \rVert_G^\ast := \max_{w \in W_G} \left\langle \rho_G, w \eta \right\rangle.\] Here \[\rho_G := \frac{1}{2} \sum_{\alpha > 0} \dim(\mathfrak{g}_\alpha) \alpha.\] We similarly define \(\lVert \eta \rVert_H^\ast\) for \(\eta \in X_\ast(T_H)\). Using these norms, we define two notions of smallness (where one should think \(\alpha = \tfrac{1}{2}\)):

To give some geometric intuition, take \(\eta \in X_\ast(T_G)\) and \(p\) prime. We can then form the corresponding Hecke operator \(\tau = \tau(\eta, p)\). What’s important about this norm is that \(\lvert \pi(H) \cap \mathop{\mathrm{supp}}(\tau) \rvert \ll \sum_{ \substack{ w \in W \\ w \eta \in X_\ast(T_H) } } p^{2 \lVert w \eta \rVert_H^\ast}\).

Back to non-concentration. What we really proved is that \(\mu(\pi(L)) = 0\) if the following condition is satisfied: each of the stabilizers \(S\) that we saw before is contained in

for all \(\varepsilon> 0\) there exists a finite set \(J_\varepsilon\in \mathcal{H}\) such that for all \(\phi\), we can find \(\tau \in J_\varepsilon\) so that \[\frac{\lVert \tau \rVert_{L^1(H)}}{\lambda_\phi} < \varepsilon.\]

Idea (Marshall): take \(\tau = \left( \sum_{p \sim P} \tau_p \right)^2\) such that each \(\tau_p\) satisfies \(\tfrac{1}{2}\)-weak smallness. That way, you hope to make the intersection small. But this is not a basic Hecke operator – you can expand it as \(\tau = \sum \tau_p^2 + \sum_{p \neq q} \tau_p \tau_q\). We may assume by some other reasons that \(\lambda_p \sim(\mathop{\mathrm{supp}}\tau_p)^{1/2}\). For the off-diagonal, we use weak smallness. For the diagonal terms, use temperedness. The problem is that in our case (dimension \(4\)), there are some very non-small groups that appear: \(H \approx \mathop{\mathrm{SO}}(1, 3) \subset \mathop{\mathrm{SO}}(1, 4)\), which is far from being \(\tfrac{1}{2}\)-tempered; it is \(\tfrac{2}{3}\)-tempered? But \(\lambda_p \geq(\mathop{\mathrm{supp}}\tau_p)^{1/2}\). The reason we find our theorem interesting is that we were nevertheless able to construct new Hecke operators such that the argument still works. This shows that the \(\tfrac{1}{2}\)-weakly small condition (used in a lot of other works) is not tight. This gives us hope for at least the following two things:

Both of these

Matthew Young, Shifted convolution problem for Siegel modular forms

(joint work with Wing Hong (Joseph) Leung; extremely recent, decided to speak on it just a week ago)

Let \(\mathbb{H}_2\) denote the Siegel upper half-space, and \(G = \mathrm{Sp}_4(\mathbb{R}) := \left\{ g \in \mathrm{GL}_4(\mathbb{R}) \mid g J g^t = J \right\}\) the symplectic group, where \(J = \left( \begin{smallmatrix} 0&-I\\ I&0 \\ \end{smallmatrix} \right)\). The group \(G\) acts on \(\mathbb{H}_2\) via fractional linear transformations: \(\gamma z =(A z + B)(C z + D)^{-1}\) for \(\gamma = \left( \begin{smallmatrix} A&B\\ C&D \\ \end{smallmatrix} \right)\).

Let \(F\) be a Siegel modular form of weight \(k\) on \(\Gamma = \mathrm{Sp}_4(\mathbb{Z})\). It admits a Fourier expansion \(F(Z) = \sum_{M \in \Lambda^+} a_F(M)\), where \(\Lambda^+ = \left\{ M \in \Lambda \mid M > 0 \right\}\). We adopt the normalization \(a_F(M) =(\det M)^{\frac{k}{2} - \frac{3}{4}} \tilde{a}_F(M)\).

Conjecture 20 (Resnikoff and Saldana, 1974). If \(4 \det M\) is a fundamental discriminant, then \(\tilde{a}_F(M) \ll(\det M)^\varepsilon\).

Conjecture 21 (Borcherer). \(\lvert a(M) \rvert^2 \doteq L(\tfrac{1}{2}, \pi \otimes \chi_d)\), where \(\pi\) lives on \(\mathrm{GSp}_4\) and \(d = \operatorname{disc}(M)\).

We’ve been imprecise here – there’s an extra average over the class group. In particular, Lindelöf does not imply Conjecture 20.

Theorem 22 (Leung, Y. 2025+). let \(Q \in \Lambda\), \(Q \neq 0\). For a “reasonable” class of test function \(W_N\) (\(N \geq 1\)), we have \[\sum_{ \substack{ M_1, M_2 \in \Lambda^+ \\ M_1 + Q = M_2 } } \tilde{a}_F(M_1) \overline{\tilde{a}_F(M_2)} W_N(M_1, M_2, Q) \ll N^{\frac{5}{2} + \varepsilon},\] where \(W_N\) is essentially supported on matrices of size \(N\).

Note that the number of matrices is essentially of size \(N^3\). The trivial bound is \(N^{3 + \varepsilon}\); this follows in particular from Conjecture 20, although we only need that on average.

The analogous problem for \(\mathrm{GL}_2\) is well-studied; other results are sparse. For instance, there has been somewhat recent progress (last 10 or 15 years) in studying, for a \(\mathrm{GL}_3\) coefficient \(\lambda_F\) and \(\mathrm{GL}_2\) coefficient \(\lambda_f\), sums like \(\sum_n \lambda_F(n) \lambda_f(n + 1)\). But we’re stuck on, e.g., \(\sum_n d_3(n) d_3(n + 1)\), which is a major bottleneck for understanding, e.g., the sixth moment of the zeta function. There’s a paper of Jaasaari–Lester–Saha where, assuming GRH, they proved QUE for Siegel modular forms of large weights \(k \rightarrow \infty\). Our method generalizes a method of DFI. The key was to find a way to interpret their method in an automorphic way, and porting that interpretation over to the land of Siegel modular forms.

Our first goal is to explain the \(\delta\)-method of DFI and indicate an automorphic way to understand it. The slogan is that the DFI \(\delta\)-method is the same as the fact that you can write the constant function as an incomplete Eisenstein series on \(\mathrm{GL}_2\): \(1 = E^{\mathrm{inc}}\) (which the speaker says he learned from (Nelson 2015)).

So let’s “recall” the DFI \(\delta\)-method. Let \(\omega : \mathbb{R} \rightarrow \mathbb{C}\), with \(\omega(0) = 0\), say of compact support. Take it to be even and normalized so that \(\sum_{n \in \mathbb{Z}} \omega(n) = 1\). The starting point of the \(\delta\)-method is the following simple fact: zero is divisible by every integer, but nonzero integers have their divisors come in pairs, since each divisor has a complementary divisor. So you can write \[\delta(n = 0) = \sum_{d \mid n} \bigl( \omega(d) - \omega\bigl(\tfrac{n}{d}\bigr) \bigr).\] Here \(d\) runs over all integers dividing \(n\). It’s convenient to write this in terms of the positive divisors, as \[2 \sum_{ \substack{ d \mid n \\ d \geq 1 } } \bigl( \omega(d) - \omega\bigl(\tfrac{n}{d}\bigr) \bigr).\] We can further write the divisibility condition using the finite Fourier transform, giving \[2 \sum_{ \substack{ d \mid n \\ d \geq 1 } } \sum_{h(d)} e_d(h n) \bigl( \omega(d) - \omega\bigl(\tfrac{n}{d}\bigr) \bigr).\] Next, writing \((h, c) = d\) and \(d = c k\), we obtain \[2 \sum_{ \substack{ d \mid n \\ d \geq 1 } } \sum_{h(d)}^\ast e_d(h n) \sum_{k \geq 1} \bigl( \omega(c k) - \omega\bigl(\tfrac{n}{c k}\bigr) \bigr).\] The sum over \(h\) is just the Ramanujan sum \(S(n, 0, c)\). One typically takes \(c \leq Q\), \(Q^2 = N\).

Next, following (Nelson 2015), let \(g :(0, \infty) \rightarrow \mathbb{C}\) be smooth, with \(g(y) \ll y^A\) as \(y \rightarrow 0\) and \(g(y) \ll 1\) as \(y \rightarrow \infty\). The incomplete Eisenstein series attached to \(g\) is \[E(z, g) := \sum_{\gamma \in \Gamma_\infty \backslash \mathrm{SL}_2(\mathbb{Z})} g(\Im \gamma z).\] By Mellin inversion and reversing orders of summation and integration, we may write the above as \[\frac{1}{2 \pi i} \int_{(2)} \tilde{g}(- s) E(z, s) \, d s.\] Pick \(\tilde{g}(- s) \zeta^\ast(2 s) \bigl( H(s) - H(1 - s) \bigr)\), with \(\zeta^\ast\) the completed zeta function and some mild conditions on \(H\), e.g., that \(H(1) - H(0) \neq 0\). Then the above becomes \[\frac{1}{2 \pi i} \int_{(2)} \bigl( H(s) - H(1 - s) \bigr) E^\ast(z, s) \, d s.\] This evaluates to the residue at \(s = 1\) plus an integral over the line \(1/2\), but the latter cancels out. Normalizing that residue to be \(1\), we obtain \[E^{\mathrm{inc}}(z) := E(z, g) = 1.\] Now, \(\delta(n = 0)\) is the \(n\)th Fourier coefficient of \(1 = E^{\mathrm{inc}}\), which may thus be written \[\frac{1}{2 \pi i} \int_{(2)} \bigl( H(s) - H(1 - s) \bigr) \bigl(\text{$n$th Fourier coefficient of $E^\ast(z, s)$}\bigr) \, d s,\] where the parenthetical Fourier coefficient is a divisor function or sum of Ramanujan sums.

The basic application of DFI is to solve the shifted convolution problem on \(\mathrm{GL}_2\). There, you obtain, for \(q \neq 0\), \[\sum_n \lambda_f(n) \lambda_f(n + q) w \bigl( \tfrac{n}{N} \bigr) \ll N^{\frac{3}{4} + \varepsilon}.\] One uses DFI, Voronoi twice (converting Ramanujan sums into Kloosterman sums), and then the Weil bound. Let’s explain how to do this same sort of thing, but using this interpretation of the \(\delta\)-method in terms of Eisenstein series (an “automorphic variant on DFI”). Let’s pick another test function \(\psi :(0, \infty) \rightarrow \mathbb{C}\), smooth, \(q \in \mathbb{Z}\), \(\psi(y) \ll y^A\) as \(y \rightarrow 0\) and \(\ll 1\) as \(y \rightarrow \infty\). Form the Poincaré series \[P_q(z, \psi) = \sum_{\gamma \in \Gamma_\infty \backslash \Gamma} e(q \Re(\gamma z)) \psi(\Im \gamma z).\] Then \[\begin{align} I &= \langle f P_q, f \rangle = \int_0^\infty \int_0^1 y^k \lvert f(z) \rvert^2 e(q x) \psi(y) \, \frac{d x \, d y}{y^2} \\ &= \sum_{m + q = n} \lambda(m) \overline{\lambda(n)} (m n)^{\frac{k - 1}{2}} \int_0^\infty y^k \exp(- 2 \pi(m + n) y) \psi(y) \, \frac{d y}{y^2}. \end{align}\] Here we think of the integral as a weight function of length \(\approx N\) if \(\psi(y)\) concentrates on \(y \asymp 1/N\). Now \[I = \langle f P_q E^{\mathrm{inc}}, f \rangle = \int_0^\infty \int_0^1 y^k \lvert f(z) \rvert^2 P_q(z, \psi) g(y) \, \frac{d x \, d y}{y^2},\] where, say, \[P_q(z, \psi) = \sum_{r \in \mathbb{Z}} a_q(r, y, \psi) e(r x),\] \[I = \sum_{ \substack{ m, n, r \\ m + r = n } } \lambda(m) m^{\frac{k - 1}{2}} \overline{\lambda(n)} n^{\frac{k - 1}{2}} \int_0^\infty \exp \left( - 2 \pi(m + n) y \right) y^{k - 1} g(y) a_q(r, y, \psi) \, \frac{d y}{y}.\] Now \(g\) decays rapidly at zero, so one should think \(y \gg 1\). The exponential has rapid decay, so one should think \(y \ll 1\). Thus, effectively, \(y \asymp 1\). Also, \(r, m, n \ll 1\), so we just need to bound the Fourier coefficients, and the bound is, for \(y \asymp 1\), \[a_q(r, y, \psi) \ll N^{\frac{3}{4} + \varepsilon}\]

There’s a standard calculation of the Poincaré series, which gives \[a_q(r, y, \psi) = \psi(y) \delta(q = r) + y \sum_{c = 1}^\infty S(r, q, c) \int_{- \infty}^\infty e\Bigl(- r x y - \frac{q x}{c^2 y(1 + x^2)}\Bigr) \, d x\] Now, \(c^2 y(1 + x^2) \asymp N\) implies \(c \leq \sqrt{N}\), so we can truncate just like in DFI. The \(x\)-integral is \(\ll \sqrt{N}/c\). One obtains \[\sum_{c \leq \sqrt{N}} \sqrt{N} \frac{\lvert S(r, q, c) \rvert}{c} \ll N^{\frac{3}{4} + \varepsilon}.\] Note that there’s no Voronoi or Bessel functions in this treatment – these are all implicit in the automorphy.

That was \(\mathrm{GL}_2\). Now we want to do the same for \(\mathrm{GSp}_4\):

  1. First, we should construct an identity \(1 = E^{\mathrm{inc}}\), which will be the analogue of DFI (although perhaps one could work out an analogue of DFI for matrices). This is harder, since \(E\) has \(12\) polar lines in \(\mathbb{C}^2\).

  2. Next, we should compute \(\langle F P_Q, F\rangle\) as a shifted convolution sum (unfolding \(P_Q\)).

  3. Finally, we unfold the other way. For this, we need a bound on the Fourier coefficients of the Poincaré series \(P_Q\). We use Kitaoka’s bound (similar to the Weil bound). The classical Weil bound \(\lvert S(m, n, c) \rvert \leq c^\varepsilon(m, n, c)^{1/2} c^{1/2}\) involves this GCD factor, and the analogous factor in Kitaoka’s work is more complicated.

Kevin Kwan, On a hybrid twisted moment for \(\operatorname{GL}(3)\) automorphic \(L\)-functions and its applications

The zeta function is defined by \(\zeta(s) = \sum_{n = 1}^\infty n^{- s} = \prod_p(1 - p^{- s})^{-1}\) for \(s = \sigma + i t\) with \(\sigma > 1\). It satisfies the functional equation \[\xi(s) := H(s) \zeta(s) = \xi(1 - s),\] where \[H(s) := \frac{1}{2} s(s - 1) \pi^{- s/2} \Gamma(s/2).\]

Motivated by the prime counting function approximation \[\sum_{p \leq x} \log p = x - \sum_\rho \frac{x^\rho}{\rho} + \mathrm{small},\] one wants to know the distribution of zeros of the Riemann zeta function. image

The basic zero counting function \[N(T) := \# \left\{ \rho = \beta + i \gamma : \zeta(\rho) = 0, \, 0 \leq \beta \leq 1, \, \lvert \gamma \rvert \leq T \right\}\] satisfies the asymptotic \[N(T) \sim \frac{T}{\pi} \log \frac{T}{2 \pi e} \sim \Delta_{[\frac{1}{2} - i T, \frac{1}{2} + i T]} \operatorname{avg} H(s).\] Hardy (1915) showed that \[N_0(T) := \# \left\{ \rho \text{ as above} : \beta = 1/2 \right\} \rightarrow \infty.\] What methods did Hardy employ to detect zeros on \(\sigma = 1/2\)?

To do better, we need to use another gift from Denmark – mollifiers, originally used by Bohr. A mollifier \(M\) only introduces more zeros, so \[(\tfrac{1}{2} - a) N_R(G) \leq(\tfrac{1}{2} - a) N_R(G M) \leq \dotsb.\] You would like to take \(M = \zeta^{-1}\), but this isn’t practical, so one instead takes something like a truncation. What Levinson did was to take \[M(s) := \sum_{h \leq y} \frac{M(h)}{h^{s + 1/2 - a}} \frac{\log(Y / h)}{\log Y} \approx \zeta(1/2 + i T)^{-1} \quad(Y = T^0).\] Assuming RH, there is an equivalent formulation due to Speiser in 1935 which says that all the zeros of the derivative of \(\zeta '(1 - s)\) are all to the left of the \(1\)-line. Later on, Levinson and Montgomery show something even more precise: if you have a zero \(\rho '_r\) for \(\zeta '\) to the right of the critical line, then you get a zero of \(\zeta\) to the left of the critical line, with control over the distances: image Levinson then takes \(f(s) =(\xi(s) + \lambda \xi '(s)) / 2\) and shows that, for some shifts \(\alpha\) and \(\beta\), \[\int_0^T \zeta(1/2 + \alpha + i t) \zeta(1/2 + \beta - i t) \Bigl( \frac{k}{h} \Bigr)^{i t} \, d t = \mathrm{MT}(h k) + \mathcal{E}(h, k)\] as \(T \rightarrow \infty\). Here \(\mathcal{E}\) is understood as an error term in the bilinear sense.

We want to do something like this more generally. We consider \[\sum_{q \sim Q} \sum_{\chi(q)}^\ast \chi(h) \overline{\chi}(k) \mathcal{M}_{\alpha, \beta}(h, k; \pi),\] where \(\pi \in \mathcal{A}_0(\mathop{\mathrm{PGL}}(3))\) and \[\label{eq:cq6w5d17ud}\tag{5} \mathcal{M}_{\alpha, \beta}(h, k;\pi) := \int_0^T L(1/2 + \alpha + i t, \pi \times \chi) L(1/2 + \beta - i t, \bar{\pi} \times \bar{\chi}) \Bigl( \frac{k}{h} \Bigr)^{i t} \, d t.\]

Theorem 23 (Conrey, Kwan, Lin, Turnage-Butterbaugh). Suppose \(\pi \in \mathcal{A}_0(\mathop{\mathrm{PGL}}(3) / \mathbb{Q})\) such that either

Suppose \(\theta \in [0,1/2)\) and \[\label{eq:cq6w5d7akn}\tag{6} Q^\varepsilon\ll T \ll Q^{\frac{1 - 2 \theta}{3 - 2 \theta} - \varepsilon},\] with \(Q \rightarrow \infty\). Let \((\lambda_h)\) be complex numbers such that

  1. \(\lambda_h \ll h^{\gamma_\theta}\) and \(\sum_{n \leq(T Q)^\theta} \frac{\lvert \lambda_h \rvert^{2 m}}{h} \ll(T Q)^\varepsilon\) for \(m = 1, 2\).

Then \[\mathcal{M}_{\alpha, \beta}(h, k; T) =(\text{expected } \mathrm{MT} \asymp T Q^2) + \mathcal{E}_{\alpha, \beta}(h, k),\] where \[\sum_{h, k \leq(T Q)^\theta} \frac{\lambda_h \overline{\lambda_k}}{\sqrt{h k}} \mathcal{E}_{\alpha, \beta}(h, k) \ll(T Q)^{7/4 + \theta / 2 + \varepsilon}.\]

Theorem 24 (CKLT-B and Farmer). For \(\pi\)’s as above and \(\varepsilon> 0\) sufficiently small, there are \(\geq(11.4 - \operatorname{O}(\varepsilon))\) percentage of zeros with \(\lvert \gamma \rvert \leq Q^\varepsilon\) for the family \(\{L(s, \pi \times \chi) : \chi(q)^\ast, \, q \sim Q\}\) that are on the critical line.

We emphasize that our results are unconditional for \(\pi = \bar{\pi}\), independent of GRC. It’s also fully unconditional for \(\mathrm{GL}(2)\) Maass forms. We note that there’s no hope to asymptotically evaluate \((5)\) when \(T\) is too small and \(Q\) is large. This motivates the assumption \((6)\). We can’t take \(T\) too large, because then we would encounter difficult shifted convolution sums. So the moral of the story is that this result is strong in the \(Q\)-aspect, while we only get “something” in the \(T\)-aspect (e.g., nothing beyond \(Q^{1/3}\)).

The idea of the proof is to use approximate functional equation and orthogonality. This gives \[\mathcal{M}(h, k, \pi) \approx T \sum_{c d \sim Q} \mu(c) \phi(d) \sum_{ \substack{ m n \ll(T Q)^3 \\ m h \equiv n k(d) \\ m h \approx n k } } \frac{\lambda_\pi(m) \lambda_{\tilde{\pi}}(n)}{\sqrt{m n}}.\] Suppose for instance that \(T \sim Q^\varepsilon\) and \(0 < 1/2 - \operatorname{O}(\varepsilon)\). We truncate and re-expand according to whether \(c > C\) or not, where \(C \approx Q^\varepsilon\). We get \[\mathcal{M}_{c > C} \approx \sum_{ \substack{ c > C \\ d \ll Q / C } }^\ast \mu(c) \sum_{\psi(d)}^\ast \psi(h) \overline{\psi}(k) \int_{t \approx T} \sum_{m n \ll(T Q / C)^{3/2}} \frac{\lambda_\pi(m) \lambda_{\tilde{\pi}}(n)}{\sqrt{m n}}.\] We then switch and re-expand and use arithmetic: writing \(m h = n k + d l\), we see that for \(m h \neq n k\), we have \(\ell \ll \frac{m h - n k}{d} \ll \frac{c}{Q}(T Q)^{3/2 + \theta}\). So the whole point of this hyperbola trick is that the length is now much shorter, by a factor of \(Q\). This enables divisor switching. Because we’re doing the aspect where \(Q \rightarrow \infty\), we can drop the condition \(m h \neq n k\). After dropping this condition, we can reexpand (mod \(\ell\)) as above. This leads to a problem in multiplicative number theory. The game is, how do you use the hybrid large sieve for the rest of the sums, making sure that the argument is independent of GRC. Previous relevant work was done by Conrey–Iwaniec–Sound. They used the bound \(\tau_3(n) \ll n^\varepsilon\) all the time. For a cuspidal automorphic representation, we only have the much weaker bound \(\lambda_\pi(n) \ll n^{5/4 + \varepsilon}\). So one needs to overcome that crutch.

Blomer, Valentin, Ashay Burungale, Philippe Michel, and Jun-Hwi Min. 2025. “Generation of Hecke Fields by Squares of Cyclotomic Twists of Modular \(L\)-Values,” March. http://arxiv.org/abs/2503.15040v2.
Nelson, Paul D. 2015. “Evaluating Modular Forms on Shimura Curves.” Math. Comp. 84 (295): 2471–2503. https://doi.org/10.1090/S0025-5718-2015-02943-3.