Abstract
We establish an asymptotic formula for the ground-state energy of the spherical pure \(( p,q) \)-spin glass model for \(p,q\ge 96\). We achieve this through understanding the concentration of the complexity of critical points with values within a region of the ground state energy. More specifically, we show that the second moment of this count coincides with the square of the first moment up to a sub-exponential factor.
Similar content being viewed by others
References
Adler, R.J., Taylor, J.E.: Random Fields and Geometry. Springer Monographs in Mathematics, Springer, New York (2007)
Agliari, E., Barra, A., Bartolucci, S., Galluzzi, A., Guerra, F., Moauro, F.: Parallel processing in immune networks. Phys. Rev. E 87, 042701 (2013)
Agliari, E., Barra, A., Galluzzi, A., Guerra, F., Moauro, F.: Multitasking associative networks. Phys. Rev. Lett. 109, 268101 (2012)
Ajanki, O.H., Erdős, L., Krüger, T.: Stability of the matrix Dyson equation and random matrices with correlations. Probab. Theory Related Fields 173(1–2), 293–373 (2019)
Amit, D.: Modeling Brain Function: The World of Attractor Neural Networks. Cambridge University Press, Cambridge (1992)
Anderson, G.W., Guionnet, A., Zeitouni, O.: An Introduction to Random Matrices. Cambridge Studies in Advanced Mathematics, vol. 118. Cambridge University Press, Cambridge (2010)
Auffinger, A., Ben Arous, G., Černý, J.: Random matrices and complexity of spin glasses. Commun. Pure Appl. Math. 66(2), 165–201 (2013)
Auffinger, A., Chen, W.-K.: Free energy and complexity of spherical bipartite models. J. Stat. Phys. 157(1), 40–59 (2014)
Azaïs, J.-M., Wschebor, M.: Level Sets and Extrema of Random Processes and Fields. Wiley, Hoboken, NJ (2009)
Baik, J., Lee, J.O.: Free energy of bipartite spherical Sherrington–Kirkpatrick model. Ann. Inst. Henri Poincaré Probab. Stat. 56(4), 2897–2934 (2020)
Barra, A., Agliari, E.: A statistical mechanics approach to autopoietic immune networks. J. Stat. Mech: Theory Exp. 2010(7), 07004 (2010)
Barra, A., Contucci, P.: Toward a quantitative approach to migrants integration. EPL (Europhys. Lett.) 89(6), 68001 (2010)
Barra, A., Genovese, G., Guerra, F., Tantari, D.: How glassy are neural networks? J. Stat. Mech: Theory Exp. 2012(07), P07009 (2012)
Barra, A., Genovese, G., Sollich, P., Tantari, D.: Phase diagram of restricted Boltzmann machines and generalized Hopfield networks with arbitrary priors. Phys. Rev. E 97, 022310 (2018)
Ben Arous, G., Bourgade, P., McKenna, B.: Exponential growth of random determinants beyond invariance (2021). arXiv:2105.05000
Ben Arous, G., Subag, E., Zeitouni, O.: Geometry and temperature chaos in mixed spherical spin glasses at low temperature: the perturbative regime. Commun. Pure Appl. Math. 73(8), 1732–1828 (2020)
Cammarota, V., Marinucci, D., Wigman, I.: On the distribution of the critical values of random spherical harmonics. J. Geom. Anal. 26(4), 3252–3324 (2016)
Cammarota, V., Wigman, I.: Fluctuations of the total number of critical points of random spherical harmonics. Stochastic Process. Appl. 127(12), 3825–3869 (2017)
Chen, W.-K.: The Aizenman–Sims–Starr scheme and Parisi formula for mixed \(p\)-spin spherical models. Electron. J. Probab. 18(94), 14 (2013)
Dembo, A., Zeitouni, O.: Large deviations techniques and applications, vol. 38 of Stochastic Modelling and Applied Probability. Springer-Verlag, Berlin (2010). Corrected reprint of the second (1998) edition
Durlauf, S.N.: How can statistical mechanics contribute to social science? Proc. Natl. Acad. Sci. 96(19), 10582–10584 (1999)
Erdős, L., Krüger, T., Schröder, D.: Random matrices with slow correlation decay. Forum Math. Sigma 7, Paper No. e8, 89 (2019)
Guerra, F.: Broken replica symmetry bounds in the mean field spin glass model. Commun. Math. Phys. 233(1), 1–12 (2003)
Ledoux, M.: Concentration of measure and logarithmic Sobolev inequalities. In: Séminaire de Probabilités, XXXIII, vol. 1709 of Lecture Notes in Math. Springer, Berlin (1999)
McKenna, B.: Complexity of bipartite spherical spin glasses (2021). arXiv:2105.05043
Nicolaescu, L.I.: Critical sets of random smooth functions on compact manifolds. Asian J. Math. 19(3), 391–432 (2015)
Nicolaescu, L.I.: Critical points of multidimensional random Fourier series: variance estimates. J. Math. Phys. 57(8), 083304, 42 (2016)
Panchenko, D.: The Parisi formula for mixed \(p\)-spin models. Ann. Probab. 42(3), 946–958 (2014)
Panchenko, D.: The free energy in a multi-species Sherrington–Kirkpatrick model. Ann. Probab. 43(6), 3494–3513 (2015)
Panchenko, D.: Free energy in the mixed \(p\)-spin models with vector spins. Ann. Probab. 46(2), 865–896 (2018)
Panchenko, D.: Free energy in the Potts spin glass. Ann. Probab. 46(2), 829–864 (2018)
Parisi, G.: A simple model for the immune network. Proc. Natl. Acad. Sci. USA 87(1), 429–33 (1990)
Simon, M.K.: Probability Distributions Involving Gaussian Random Variables: A Handbook for Engineers. Scientists and Mathematicians, Springer-Verlag, Berlin, Heidelberg (2006)
Subag, E.: The complexity of spherical \(p\)-spin models-a second moment approach. Ann. Probab. 45(5), 3385–3450 (2017)
Talagrand, M.: Free energy of the spherical mean field model. Probab. Theory Related Fields 134(3), 339–382 (2006)
Talagrand, M.: The Parisi formula. Ann. Math. (2) 163(1), 221–263 (2006)
Acknowledgements
The author would also like to thank their advisor, Antonio Auffinger, for introducing them to this problem, and for many comments on earlier versions of this manuscript.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by S. Chatterjee.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Research partially supported by NSF Grants DMS-2202720, DMS-1653552 and DMS-1502632.
Appendices
Covariance Computations and Conditional Densities
In this appendix we will study the covariance structure of the Gaussian vector
after which we will investigate some of its conditional density functions.
We will be interested specifically in studying the case where the orthonormal frame field \((E_i)_{i=1}^{N-2}\) is induced by a pullback of a choice of an orthonormal frame field on each component of the product. That is, let us write \(E^i=E_{i+N_1-1}\), and assume that \((E_i)_{i=1}^{N_1-1}\) is obtained from an orthonormal frame field on the factor \(S^{N_1-1}\), and similarly that \((E^i)_{i=1}^{N_2-1}\) is obtained from an orthogonal frame field on the factor \(S^{N_2-1}\). We observe that in this case, the action of \(E^i\) commutes with \(E_j\) for each i, j.
With \(\delta _{ij}\) denoting the standard Kronecker \(\delta \), we denote \(\delta _{ijk}=\delta _{ij}\delta _{jk}\), \(\delta _{ij\ne 1}=\delta _{ij}(1-\delta _{i1})\), etc. In addition, we employ the shorthand \(x_*=\sqrt{1-x^2}\).
Lemma 14
For \(p,q\ge 2\), \(0<\gamma <1\) and \(r,t\in [-1,1]\) there exists an orthogonal frame field E, such that:
Proof
The proof will proceed similarly to Lemma 30 of [34]. Let us denote \(r_1=r\) and \(r_2=t\). Let \(P_{{{\textbf {n}}},N}:S^{N-1}\rightarrow {\mathbb {R}}^{N-1}\) denote projection away from the first coordinate, let \(\theta _i\in [-\pi /2,\pi /2]\) denote the angles such that \(\sin (\theta _i)=r_i\), and let \(R_{\varphi ,N}:S^{N-1}\rightarrow S^{N-1}\) denote the rotation mapping
For \(i=1,2\), take a neighborhood \(U_i\) of \({{\textbf {n}}}_{N_i}\) in \(S^{N_i-1}\) such that restriction of \(P_{{{\textbf {n}}},N_i}\) is a diffeomorphism onto its image. Similarly take a neighborhood \(V_i\) of \({{\textbf {n}}}_{N_i}(r_i)\) in \(S^{N_i-1}\) such that \(P_{{{\textbf {n}}},N_i}\circ R_{-\theta _i,N_i}\) is a diffeomorphism onto its image. We denote these images as \({\bar{U}}_i\) and \({\bar{V}}_i\), respectively. We define functions \({\bar{h}}_1:{\bar{U}}_1\times {\bar{U}}_2\rightarrow {\mathbb {R}}\) and \({\bar{h}}_2:{\bar{V}}_1\times {\bar{V}}_2\rightarrow {\mathbb {R}}\) by
In the proof of Lemma 30 of [34] (and more specifically in their ft. 5) they show that for each \(N\ge 2\) and \(\theta \in [-\pi /2, \pi /2]\), there is an orthonormal frame field, \(E^{N,\theta }=(E_i^{N,\theta })_{i=1}^{N-1}\), on \(S^{N-1}\) with the following property: for any smooth function \(f:S^{N-1}\rightarrow {\mathbb {R}}\) if we denote
then for \(1\le i,j\le N-1\) we have that
where \(\partial _i\) denotes the i-th standard Euclidean partial derivative. We define our orthogonal frame field E on \(S^{N_1-1}\times S^{N_2-1}\) by letting \((E_i)_{i=1}^{N_1-1}=(E_i^{N_1,\theta _1})_{i=1}^{N_1-1}\), where \(E_i^{N_1,\theta _1}\) acts on the coordinates of the first sphere in the product, and \((E^i)_{i=1}^{N_2-1}=(E_i^{N_2,\theta _2})_{i=1}^{N_2-1}\), where \(E_i^{N_2,\theta _2}\) acts on the coordinates of the second sphere. Let us write the i-th Euclidean coordinate for \({\bar{U}}_1\subseteq {\mathbb {R}}^{N_1-1}\) as \(x_i\) and the i-th Euclidean coordinate for \({\bar{U}}_2\subseteq {\mathbb {R}}^{N_2-1}\) as \(y_i\). Then employing (267) and (268) we see that for \(1\le i,j\le N_1-1\) and \(1\le k,l\le N_2-1\)
A similar relationship holds for between the derivatives of \(h_N\) at \({{\textbf {n}}}(r,t)\) and \({\bar{h}}_2\) at 0.
If we now define
and let \(C(x,y,z,w):={\mathbb {E}}[{\bar{h}}_1(x,y){\bar{h}}_2(z,w)]\), then for \(x,z\in {\mathbb {R}}^{N_1-1}\) and \(y,w\in {\mathbb {R}}^{N_2-1}\), we have
We recall (5.5.4) of [1], which states that for an arbitrary centered Gaussian field on \({\mathbb {R}}^n\), h, with smooth covariance function, C, we have that
Routine calculation using these formulas now yields the desired results. \(\square \)
From this result we observe the following corollary.
Corollary 2
Let us take \(r,t\in [-1,1]\) and the choice of frame field in Lemma 14. Then all the following random variables are independent.
-
\(E_i E_j h_N({{\textbf {n}}})+p\delta _{ij}h_N({{\textbf {n}}})\) for \(1<i\le j\le N_1-1\).
-
\(E^i E^j h_N({{\textbf {n}}})+q\delta _{ij}h_N({{\textbf {n}}})\) for \(1<i\le j\le N_2-1\).
-
\(E_i E^j h_N({{\textbf {n}}})\) for \(1<i\le N_1-1,1<j\le N_2-1\).
-
\((E_{1}E_i h_N({{\textbf {n}}}),E^1 E_i h_N({{\textbf {n}}}),E_i h_N({{\textbf {n}}}),E_i h_N({{\textbf {n}}}(r,t)))\) for any \(1<i\le N_1-1\).
-
\((E_{1}E^i h_N({{\textbf {n}}}),E^1 E^i h_N({{\textbf {n}}}),E^i h_N({{\textbf {n}}}),E^i h_N({{\textbf {n}}}(r,t)))\) for any \(1<i\le N_2-1\).
-
\((E_{1}E_1h_N({{\textbf {n}}}),E_{1}E^1h_N({{\textbf {n}}}),E^{1}E^1h_N({{\textbf {n}}}),h_N({{\textbf {n}}}),h_N({{\textbf {n}}}(r,t))\),
\(E_1h_N({{\textbf {n}}}),E^1h_N({{\textbf {n}}}),E_1h_N({{\textbf {n}}}(r,t)),E^1h_N({{\textbf {n}}}(r,t)))\).
We will now turn our attention to the computation of the relevant conditional densities, beginning with that of \((\nabla h_N({{\textbf {n}}}),\nabla h_N({{\textbf {n}}}(r,t)))\). In what follows we will repeatedly use the formula for the conditional law of Gaussian distributions (see for example, (1.2.7) and (1.2.8) of [1]). We observe that for \(1<i<N_1-1\) and \(1<j<N_2-1\)
From these we see that the vectors \((E_i h_N({{\textbf {n}}}), E_i h_N ({{\textbf {n}}}(r,t)))\) and \((E^j h_N({{\textbf {n}}}), E^j h_N({{\textbf {n}}}(r,t)))\) are non-degenerate unless \(|r|=|t|=1\). The remaining entries are more complicated. We may compute that
so that if we denote the covariance matrix of the vector
as \(\varSigma _L(r,t)\) then we have that
We note for convenience that
We observe that when \(\varSigma _L(r,t)\) is invertible, the density of the vector (278) at zero is
Proof of Lemma 9 and Lemma 10
By Lemma 17 below, for any \(r,t\in (-1,1)\) we have that the vector
is non-degenerate. Equation (143) now follows from (274) and (275) and the formula for the conditional density of a Gaussian random variable. We will use the notation
We define
Then Lemma 17 guarantees that \(\varSigma _U(r,t)\) is strictly positive-definite, and the remaining claims follow from Lemma 17 and the formulas for the conditional law of Gaussian distributions. Together these results complete the proof of Lemma 9.
The proof of Lemma 10 follows from Lemma 14 and Corollary 2 and the same conditional formulas. \(\square \)
The next result shows that the only points where \(\det (\varSigma _L(r,t))=0\) have \(|r|=|t|=1\).
Lemma 15
For \(p,q\ge 5\), there is a constant \(C>0\), such that for \(r,t\in [-1,1]\), we have that
Proof
Appealing to Lemma 17 as before, for \(r,t\in (-1,1)\) we have that the vector
is non-degenerate, so that in particular \(\det (\varSigma _L(r,t))>0\). From this and continuity, we also see that \(\det (\varSigma _L(r,t))\ge 0\), so that, in particular, it suffices to show that there is \(c>0\) such that
We observe that
We note that for \(|r|<1\), we have by the AM-GM inequality that
so in particular, \((p-1)r^{p-2}(1-r^2)<1\) and so \(1+r^{p-2}(pr^2-(p-1))\ge 1-(p-1)r^{p-2}(1-r^2)>0\). As \(|r|<1\), we more easily see that \(1-r^{p-2}(pr^2-(p-1))\ge 1-r^{p-2}>0\) and that \(|r|^p<1\). In particular, we see that if \(|r|<1\) then
Similarly we see that if \(|t|<1\) then \(\det (\varSigma _L(\pm 1,t)>0\). Thus what remains is to study the order of vanishing of \(\det (\varSigma _L(r,t))\) around the points \(|r|=|t|=1\). We observe that \(\det (\varSigma _L(r,t))\) is even in both r and t, so with the prior positivity results, it is sufficient to show that
where the limit here (and below) is taken with \(r,t\in (-1,1)\). Let us denote \(\delta =(2-r^2-t^2)\). Then for \(r,t\in (-1,1)\) sufficiently close to (1, 1), we note that
so that we have that
Finally a long but direct computation yields that
Let us denote the matrix on the right-hand side as \(H_L(p,q)\). We see that by Taylor’s theorem
On the other hand for \(p,q\ge 2\) all entries of \(H_L(p,q)\) are positive, so for small enough \(c>0\) and \(r,t\in (-1,1)\)
Combining this with the observation that
then completes the proof. \(\square \)
Remark 4
By Lemma 15, (274), (275), and Corollary 2 we see that for \(r,t\in [-1,1]\), such that either \(|r|<1\) or \(|t|<1\), we have that \((\nabla h_N({{\textbf {n}}}),\nabla h_N({{\textbf {n}}}(r,t)))\) is a non-degenerate Gaussian random vector. As the covariance function for \(h_N\) is isotropic, we see by applying a rotation, that as long as either \(N^{-1}_1|(\sigma ,\sigma ')|<1\) or \(N^{-1}_2|(\tau ,\tau ')|<1\), the law of
is non-degenerate.
We will also need a result that controls the rate at which the entries of the Hessian vanish as \(r,t\rightarrow 1\).
Lemma 16
There is a fixed constant \(C>0\), only dependent on p, q, such that for \(1\le i\le N_1\), we have that
and for \(1\le i\le N_2\), we have that
Proof
We only show (300), the case of (301) being similar. We also note that by symmetry and continuity, we may assume that \(r,t\in (0,1)\). As the quantity on the right is positive, we may additionally assume that \(r,t\in (1-\epsilon ,1)\) for any fixed \(\epsilon >0\).
Let \({\bar{h}}_1\) and \({\bar{h}}_2\) be as in Lemma 14, and let us denote \(v_{N}(s)=(\sqrt{1-s^2},0,\dots 0)\in {\mathbb {R}}^{N-1}\) and \(v(r,t)=(v_{N_1}(r),v_{N_2}(t))\). As both \(\nabla {\bar{h}}_2(0)\) and \(\nabla {\bar{h}}_1(v(r,t))\) are coordinate expressions of the gradient with respect to (potentially different) orthonormal frames, they are related at a fixed point by an invertible matrix. In particular, see that the linear relation \(\nabla h_N({{\textbf {n}}}(r,t))=\nabla {\bar{h}}_2(0)=0\) is equivalent to the linear relation \(\nabla {\bar{h}}_1(v(r,t))=0\). If we denote \({\bar{h}}_1={\bar{g}}\) we see then that
With this notation, we see from the proof of Lemma 14 that we may rewrite (300) as
We will write \((x,y)\in {\mathbb {R}}^{N_1-1}\times {\mathbb {R}}^{N_2-1}\) to denote the standard Euclidean coordinates. Let us denote, for \(1\le i\le N_1\)
We observe that by Taylor’s Theorem, we have that
Observing that \(\Vert (\sqrt{1-r^2},\sqrt{1-t^2})\Vert ^2=(2-r^2-t^2)\) we thus see by (302) that this implies that
Employing that \(\Vert H_{u,v}^i\Vert \le \Vert H_{u,v}^i\Vert _{HS}\), and that conditioning on a Gaussian vector may only decrease its variance, we see that
As the covariance of the entries of \(H_{u,v}^i\) are clearly bounded (uniformly in u, v and N), we see that the right-hand side of (309) is bounded by a fixed constant by the Borell-TIS inequality (see Theorem 2.1.1 of [1]). In view of (303), this completes the proof. \(\square \)
Finally, we will end this section by computing a useful expression involving \(\varSigma _U(r,t)\) that we will use above. We will fix \(r,t\in (-1,1)\) for the remainder of the section. We first note that by Corollary 2 and (284) that
To simplify this further, we observe that by direct computation from Lemma 14 the vector
and the vector
are independent. The covariance matrix of the vector (312) is given by
We observe that by the above independence we see that
We now proceed to compute this quantity. We first note that by direct computation, we have that
We will also need a term from the adjugate matrix, namely
Together these yield that
so altogether, we see that
Proof of Lemma 8
In this section, we provide a proof of Lemma 8. The proof will be almost identical to that of Lemma 4 of [34], which may be immediately adapted to this case if we prove the following result.
Lemma 17
For \(p,q\ge 5\), and \(r,t\in (-1,1)\), we have that the Gaussian vector
is non-degenerate, up to degeneracies required by symmetries of the Hessian. That is, the vector is non-degenerate if one only takes elements of the Hessians above and on the main diagonal.
Proof of Lemma 8
One may proceed almost identically to the proof of Lemma 4 of [34], replacing their Lemma 32 with our Lemma 17. \(\square \)
The proof of Lemma 17 will follow from two lemmas, both proven at the end of this section. To state the first of these, we first recall the (normalized) spherical \(\ell \)-spin glass model (see (1.1) and (4.1) of [34]). For \(\ell \ge 1\) and \(N\ge 2\), the (normalized) \(\ell \)-spin glass model, is a smooth centered Gaussian random field \(h_{N,\ell }:S^{N-1}\rightarrow {\mathbb {R}}\), with covariance
We see that this covariance is related to that of the (p, q)-spin glass model in that
For \(r\in (-1,1)\), let us denote \({{\textbf {n}}}(r)={{\textbf {n}}}_N(r)=(r,\sqrt{1-r^2},0,\dots ,0)\) and \({{\textbf {n}}}={{\textbf {n}}}(1)\). Furthermore, let \((E_i)_{i=1}^{N-1}\) denote the choice of (piece-wise) smooth orthonormal frame field on \(S^{N-1}\) defined in Lemma 30 of [34], and define \((\nabla h_\ell ,\nabla ^2 h_\ell )\) as above.
Lemma 18
For \(\ell \ge 5\), and \(r\in (-1,1)\), the Gaussian vector
is non-degenerate, up to degeneracies required by symmetries of the Hessian.
Remark 5
This is a similar to (and reliant on) Lemma 32 of [34], which establishes that for \(\ell \ge 3\) and \(r\in (-1,1)\), the Gaussian vector
is non-degenerate, up to degeneracies required by symmetries of the Hessian.
We observe that the E defined in the proof of Lemma 14 is such that \((E_i)_{i=1}^{N_1-1}\) and \((E^i)_{i=N_1}^{N-2}\) are given by the extension to \(S^{N_1-1}\times S^{N_2-1}\) of derivations acting only \(S^{N_1-1}\) and \(S^{N_2-1}\), respectively. Moreover, by construction, these are chosen to so that their restrictions coincide with the E defined above in Lemma 18 with \(\ell =p\), \(N=N_1\), and \(r=r\) and with \(\ell =q\), \(N=N_2\) and \(r=t\), respectively. In particular, we see that Lemma 17 follows from Lemma 18 and Lemma 19 below.
Lemma 19
For \(1\le i\le 2\), let \(h_i\) denote a smooth centered Gaussian field defined on an open subset \(U_i\subseteq {\mathbb {R}}^{n_i}\), with smooth covariance function \(f_i\). Let h denote the smooth centered Gaussian field defined on \(U_1\times U_2\) with covariance function \(f_1 f_2\). For some \(\ell \ge 1\) and \(1\le i\le 2\), let us choose some sequence of points \(r_1^i,\dots r_\ell ^i\in U_i\), and let us denote \(r_{k,l}=(r_k^1,r_l^2)\). Then if we have, for both \(l=1,2\), that the Gaussian vector comprised of entries
is non-degenerate up to degeneracies required by symmetries of the Hessian, then the Gaussian vector comprised of entries
is non-degenerate up to degeneracies required by symmetries of the Hessian.
We now complete this section by giving the proofs of Lemmas 18 and 19.
Proof of Lemma 18
For the duration of this proof we will denote \(h_N:=h_{N,\ell }\). To establish the desired claim, it is sufficient to show both that the vector
is non-degenerate, and that the law of \((\nabla ^2 h_{N}({{\textbf {n}}}), \nabla ^2 h_{N}({{\textbf {n}}}(r)))\) conditioned on the event
is non-degenerate.
For the ease of the reader, we will begin by recalling all the results of [34] that we will use. For clarity, their (f, p) coincides with our \((h_N,\ell )\). In their Lemma 12 they show that the law of \((\nabla h_{N}({{\textbf {n}}}),\nabla h_{N}({{\textbf {n}}}(r)))\) is non-degenerate, and that the covariance matrix of \((h_{N}({{\textbf {n}}}),h_{N}({{\textbf {n}}}(r)))\) conditional on the event \((\nabla h_{N}({{\textbf {n}}})=\nabla h_{N}({{\textbf {n}}}(r))=0)\) is given by a matrix \(\varSigma _U(r)\), which they show is invertible in Remark 31. Together these show that (330) is non-degenerate. Now conditional on (331), their Lemma 13 shows that the only non-trivial correlations between entries of \((\nabla ^2h_N ({{\textbf {n}}}),\nabla ^2h_N ({{\textbf {n}}}(r)))\) are between \((E_{i} E_j h_N({{\textbf {n}}}),E_{i} E_j h_N({{\textbf {n}}}(r)))\), and those enforced by symmetry of the Hessian. Thus we are reduced to showing that the conditional law of \((E_{i} E_j h_N({{\textbf {n}}}),E_{i} E_j h_N({{\textbf {n}}}(r)))\) is non-degenerate for each \(1\le i\le j\le N-1\). The case of \(i,j<N-1\) is clear from their description in item (2) of Lemma 13, and the case of \(i<N-1\) and \(j=N-1\) follows from their description in item (3) and their proof of Lemma 32, where they show that \(\varSigma _Z(r)\) is invertible. The remaining case is \(i=j=N-1\), where the conditional covariance matrix is proportional to the matrix \(\varSigma _Q(r)\) they define in (10.2).
In principle, one should be able to derive the invertibility of \(\varSigma _Q(r)\) directly from this expression, but due to the technicality of this expression, we instead employ a more abstract argument. To perform this, we need only observe that the expression of \(\varSigma _Q(r)\) is N-independent, so that to show that \(\varSigma _Q(r)\) is invertible, it is thus sufficient to show that the conditional law of \((E_{N-1}E_{N-1}h_N({{\textbf {n}}}),E_{N-1}E_{N-1}h_N({{\textbf {n}}}(r)))\) is non-degenerate in the case of \(N=2\). Henceforth, we assume that \(N=2\) and omit N from the notation.
We observe that \(h_\ell \) coincides in law with the Gaussian field defined for \((x,y)\in S^1\) by
where here \((a_i)_{i=0}^{\ell }\) are independent centered Gaussian random variables with \({\mathbb {E}}[a_i^2]=\left( {\begin{array}{c}\ell \\ i\end{array}}\right) \). We observe that by continuity of the derivatives of f as a function of \((a_i)_{i=1}^\ell \), if the matrix \(\varSigma _Q(r)\) is degenerate, any (deterministic) homogeneous polynomial of degree-\(\ell \) on \(S^1\) satisfying the linear relation (331) must satisfies a fixed non-trivial linear relationship between the elements of \((E_{1} E_1 f({{\textbf {n}}}),E_1 E_1 f({{\textbf {n}}}(r)))\). On the other hand, the polynomials
satisfy (330) but in addition have
Thus, no such relationship is possible, which shows that the law of
is non-degenerate when \(N=2\) and thus in general. \(\square \)
Proof of Lemma 19
This proof will consist of showing that the covariance matrix of (329), after removing duplicate entries of the Hessians, is a submatrix of the Kronecker product of the covariance matrices of (328), where again we have removed duplicate entries. Indeed, as the Kronecker product of two strictly positive definite matrices is strictly positive definite, the lemma then follows from the interlacing property for the eigenvalues of a submatrix of a symmetric matrix.
To demonstrate that one is a submatrix of the other, we will make repeated use of the identity (273) above, which expresses the covariance between derivatives of a Gaussian field in terms of the derivatives of the covariance function. For ease of notation, we introduce the vector-valued operator \(\nabla ^{U}\), which is defined for a function f in neighborhood of \(x\in {\mathbb {R}}^N\) as
considered as an \(N(N+1)/2\)-dimensional vector with the lexicographic ordering on (i, j). Then for \(l=1,2\), we see from (273) that covariance of the vector \((h_l,\nabla h_l,\nabla ^{U} h_l)\) evaluated at two points \(x,y\in {\mathbb {R}}^{n_l}\) is given by
where here \(\nabla _1 f_l\) and \(\nabla _{2} f_l\) denote the gradient of \(f_l:U_l\times U_l\rightarrow {\mathbb {R}}\) taken with respect to the first and second copy of \(U_l\), respectively (and similarly for \(\nabla ^{U}\)) and where the matrix indices such that the index of \(\nabla _2\) and \(\nabla ^{U}_2\) give the row number and the index of \(\nabla _1\) and \(\nabla ^{U}_1\) give the column number. With this notation we see that the covariance of the matrix of (328), with repeated entries removed, is up to reordering given by
To specify the covariance matrix of (328), we note that covariance function of h is given by \(f_1f_2:U_1\times U_2\times U_1\times U_2\rightarrow {\mathbb {R}}\), or explicitly, for \((x,y,z,w)\in U_1\times U_2\times U_1\times U_2\) we have that
We are interested in the covariance matrix of the vector \((h,\nabla h,\nabla ^{U} h)\) at two points (x, y), \((z,w)\in U_1\times U_2\). We will denote by \((\nabla _I,\nabla _{II})\) the decomposition of the gradient on \(U_1\times U_2\) into the first and second factor, and similarly for \((\nabla _{I}^{2,U},\nabla _{II}^{2,U})\). Using this notation, we will also reorder and write the vector \((h,\nabla h,\nabla ^{U} h)\) as \((h,\nabla _I h,\nabla ^{U}_I h,\nabla _{II} \)\( h,\nabla _{I} \nabla _{II} h,\nabla ^{U}_{II} h)\). With respect to this decomposition, we may write the covariance matrix in terms of the induced 6-by-6 block decomposition as
where here \(f_1=f_1(x,z)\), \(f_2=f_2(y,w)\) (and similarly for their derivatives), \(\nabla _{12}=\nabla _1\nabla _2\) and \(\nabla _i f_l\) is given by its expression above. The covariance matrix of (329), with degenerate entries removed, is then given by up to reordering by
Now the matrix \(\varSigma ^1(x,z)\otimes \varSigma ^2(y,w)\), where \(\otimes \) denotes the Kronecker product, may be decomposed into blocks labeled by (i, j) with \(1\le i,j\le 3\), by employing the block decomposition of (337). We observe then that \(\varSigma (x,y,z,w)\) is equivalent to the submatrix of \(\varSigma _1(x,z)\otimes \varSigma _2(y,w)\), consisting of blocks with indices (i, j) with \(i+j\le 3\). As the choice of indices does not depend on the choice of (x, y, z, w), we see as well that \(\varSigma \) is a submatrix of \(\varSigma ^1\otimes \varSigma ^2\). \(\square \)
Proof of Eqs. 3 and 4
In this section, we will show that formulas (3) and (4) coincide with their counterparts in Theorem 2.1 of [25]. In particular, we need to relate our \(\mu _{p,q,\gamma }\) to the “\(\mu _{\infty }(u)\)” defined in Remark 2.2 and 2.3 of [25], and in addition, we must relate the quantity “\(E_{\infty }(p,q,\gamma )\)”, defined in their Theorem 2.1 and Remark 2.3 to our \(E_{p,q,\gamma ;\infty }\). To avoid confusion, let us denote these quantities as \(\mu _{p,q,\gamma ,u}^{Mc}\) and \(E_{p,q,\gamma ;\infty }^{Mc}\), respectively. Specifically, to verify the desired claims of (3) and (4) we need to show that
as well as verifying that the constructions given in Sect. 1.1 are indeed possible. In particular, given (343) we show that the complexity function \(\varSigma _{p,q,\gamma }\), given in (14) and (15), coincides with the function given in [25].
Before proceeding, we will explain the reason for the difference between our expressions and summarize the method to demonstrate (343). The difference is due to our rescaling of the rows and columns in the Hessian appearing in the Kac-Rice formula before passing to the limiting spectral measure. This trick, specific to the pure case, allows us to realize all the limiting empirical measures as translations of a single fixed measure. On the other hand, while the effect of our rescaling has an obvious effect on the determinant, it is less clear what effect it has on the limiting spectral measure, as there appears to be no direct way to compare the solution to our (13) to the solutions of (2.6) in [25]. Instead, we will use that this rescaling does not affect any of the proofs present in [25], and so we obtain an alternative expression for the complexity, which, when combined with the expression in for this in [25] gives (343).
To begin, we recall some results used in the proof of Theorem 2.1 in [25]. In particular, we need the following results, which follow from their application of the Kac-Rice formula in Lemma 3.2 (and noting that their \(\alpha _1=\alpha _2=0\) in the pure case, so that \(H_N(u)\) only depends on \(u_0\)): for nice \(B\subseteq {\mathbb {R}}\),
where here \(H_N(u)\) is an \((N-2)\)-by-\((N-2)\) Gaussian symmetric random matrix with independent entries, satisfying
Following their proof of Theorem 2.1 we see that
Let us define the matrix
and define
If we denote \(H^D_N=H^D_N(0)\), we note that \(H^D_N(u)\) coincides with \(H_N^D-uI\) in law. We also note that
Noting as well that the index is invariant under congruence transformations, we obtain the following result, which we formulate as a self-contained lemma for usage above.
Lemma 20
For \(p,q\ge 2\), \(0<\gamma <1\), and nice \(B\subseteq {\mathbb {R}}\) we have that
where \(H^{D}_N\) is a symmetric random matrix with independent centered Gaussian entries with covariances for \(1\le i,j\le N-2\) satisfying
Now, proceeding through their Section 3, replacing \(H_N(u)\) by \(H_N^D(u)\), rescaling terms as necessary, the proofs in Section 3 of [25] give the following result.
Lemma 21
Fix \(p,q\ge 2\) and \(0<\gamma <1\). Then there exists a unique compact-supported probability measure with continuous bounded density on \({\mathbb {R}}\), \(\mu _{p,q,\gamma }\), specified by the procedure in (13). Moreover, for nice \(B\subseteq {\mathbb {R}}\)
Now employing this, and comparing the terms in Lemma 20 and equations (351) and (353) we may conclude (343).
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Kivimae, P. The Ground State Energy and Concentration of Complexity in Spherical Bipartite Models. Commun. Math. Phys. 403, 37–81 (2023). https://doi.org/10.1007/s00220-023-04733-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00220-023-04733-6