Block diagonalizing ultrametric matrices

1 downloads 0 Views 231KB Size Report
between replicas, allows us to bring all ultrametric matrices to a block .... these at most two can be different, fαβγ is, in fact, a function of only two variables, e.g. of.
arXiv:cond-mat/9409050v2 14 Sep 1994

BLOCK DIAGONALIZING ULTRAMETRIC MATRICES

T. Temesv´ ari* Institute for Theoretical Physics, E¨otv¨os University, H-1088 Puskin u 5-7, Budapest, Hungary C. De Dominicis Service de Physique Th´eorique, CEA, Saclay F-91191 Gif-sur-Yvette Cedex, France I. Kondor Bolyai College, E¨otv¨os University, H-1145 Amerikai u ´t 96, Budapest, Hungary

Abstract The problem of diagonalizing a class of complicated matrices, to be called ultrametric matrices, is investigated. These matrices appear at various stages in the description of disordered systems with many equilibrium phases by the technique of replica symmetry breaking. The residual symmetry, remaining after the breaking of permutation symmetry between replicas, allows us to bring all ultrametric matrices to a block diagonal form by a common similarity transformation. A large number of these blocks are, in fact, of size 1 × 1, i.e. in a vast sector the transformation actually diagonalizes the matrix. In the other sectors we end up with blocks of size (R + 1) × (R + 1) where R is the number of replica symmetry breaking steps. These blocks cannot be further reduced without giving more information, in addition to ultrametric symmetry, about the matrix. Similar results for the inverse of a generic ultrametric matrix are also derived. PACS classification numbers: 75.10.Nr, 05.50.+q. * E-mail:[email protected]

1. Introduction Low temperature disordered systems often possess many equilibrium phases. The technique of replica symmetry breaking (RSB) provides a theoretical framework in which these systems can be described analytically , starting from a microscopic basis. Discovered and developed in the theory of spin glasses [1], RSB has recently penetrated into a number of other problems, including the theory of random manifolds [2-4], random field problems [5,6], protein folding [7-9], vortex pinning [10], etc. In each of these theories randomness is handled via the replica trick, and the multitude of equilibrium phases is captured by breaking the permutation symmetry between the replicas. As always, symmetry breaking means that the low temperature solutions realize a particular subgroup of the underlying symmetry group of the theory, here of the permutation group. The proper choice of the subgroup proved to be a highly nontrivial task in the case of RSB. The succesful Ansatz for the symmetry breaking pattern, proposed by Parisi originally in the context of spin glasses, turned out to embody a particular, hierarchical organization of the equilibrium phases, usually referred to as ultrametricity [1]. The corresponding subgroup determines the structure not only of the order parameter, but also of all other quantities in the theory, like self-energies, propagators, etc. The structure imposed by this residual symmetry on quantities depending on two replica indices is by now widely known. The algebra of such quantities has been worked out by Parisi [11] with further results, most notably on the inversion problem, added by M´ezard and Parisi [2]. At a certain stage of the development of RSB theories, however, one has to face also more complicated objects, depending on three or four replica indices. The structure of these is much harder to grasp and their algebra is much more involved than that of the two-index quantities. Our purpose here is to analyse and exploit the structure imposed by ultametricity on four-index quantities. For reasons to become clear shortly, we shall call them ultrametric matrices and will be concerned, in particular, with their diagonalisation and inversion. In order to keep the full generality of the results and thereby guarantee their applicability in any RSB theory, we shall not assume any properties other than those imposed by ultrametricity on these matrices. This way we separate the analysis of the purely geometric aspects of RSB theories, which are common to all of them, from the treatment of other properties which are determined by more specific details of the particular systems. Also, we shall keep the number n of replicas a positive integer throughout this paper. The replica limit n → 0 is, of course, the most essential step of the replica method. It is also the source of mathematical ambiguities. The analysis of the consequences of ultrametric symmetry, however, does not depend on n, therefore we found it useful to keep it finite. This way our analysis belongs to the realm of well-established mathematics and the analytic continuation in n can be carried out at the latest stage, on the final results. The number R of replica symmetry breaking steps will also be considered a generic integer. Thus our results will be applicable in situations where only a single RSB step is needed, as well as in the case of full-fledged RSB with R → ∞. The results for this

”continuous” case (R → ∞, n → 0), derived by a completely different method, will be published in [12]. Although almost trivial in principle, our analysis will, inevitably, be very complicated in actual details. It is clearly impossible to reproduce the often very lenghty calculations here, and we shall have to use the phrase ”it can be shown” frequently. What we mean at these points is that one can reproduce the results easier than to follow the lengthy proofs. A good strategy is to work out a simple special case (like that with R = 1) first; the induction is easy to spot in most cases. The plan of the paper is the following: In Sec. 2 the definition of ultrametric matrices is presented together with the classification of their different matrix elements. Sec. 3 contains a detailed analysis of the non-orthogonal basis vectors of a similarity transformation that brings all ultrametric matrices to a block diagonal form. In Sec. 4 this block diagonal form is expressed through some ”kernels”, which facilitate the eigenvalue and inversion problem greatly. A complete list of matrix components versus kernels is also included in this section. Some technicalities are relegated to the Appendix. 2. Definition of a generic ultrametric matrix For the sake of definiteness we present our analysis in the language of spin glasses, the extension to other replica symmetry breaking theories is merely a matter of notation. The replica method yields the free-energy F of a long-range spin glass in the form of a functional, depending on a set qαβ of order parameters: F = F (qαβ ). The replica indices α, β take integer values: α, β = 1, 2, . . . , n (for our present analysis the replica limit n → 0 need not be considered here). The order parameters are symmetric: qαβ = qβα , and (for Ising spins) the diagonal components are zero: qαα = 0. The number of independent order parameter components is thus 12 n(n − 1). The free energy is independent of the labeling of the replicas, so F must be constructed from the algebraic invariants of the permutation group of n objects. Examples of such invariant combinations are:

X αβ

X

qαβ ,

αβ

4 qαβ ,

T r q4 ,

T r q2 = X

αβγ

X αβ

2 2 qαβ qβγ ,

2 qαβ ,

T r q3 = X

X

qαβ qβγ qγα ,

αβγ

2 qαβ qαγ qγδ qδα ,

etc.

αβγδ

The stationary values of the order parameter components are determined by the equation of state: ∂q∂Fαβ = 0. There are 12 n(n − 1) such equations. Depending on the parameters in F these equations may have solutions that are indentically zero, qαβ = 0, ∀α, β, or Kr that have non-zero but identical off-diagonal components qαβ = q(1 − δα,β ), or non-zero off-diagonal components that depend on the pair (α, β) of replica indices. The solutions of the last kind are said to be replica symmetry breaking (RSB) solutions and these are the ones that describe situations where many equilibrium states exist. Led by a number of formal considerations that later turned out to embody the ultrametric organization of these states, Parisi proposed a, by now standard, parametrization for the RSB solutions which we briefly recapitulate in order to fix notations.

Firstly assume that n is not only an integer, but a very large one, with a large number of proper divisors. Let p1 be one of these, itself a large number with many divisors, one of them p2 , etc. up to pR . It is useful to rename n as p0 , and add pR+1 ≡ 1 to the other end of the series. Now the n replicas are divided into n/p1 boxes each containing p1 replicas. The contents of each box are further subdivided into p1 /p2 smaller boxes with p2 replicas in each etc., down to the smallest boxes with pR replicas. The RSB solutions are supposed to be invariant w.r.t. the permutations of replicas inside each of the smallest boxes of size pR , and also w.r.t. the permutations of the size pk+1 boxes inside each of the size pk boxes for any k = 0, 1, . . . , R − 1. Evidently, these permutations form a subgroup of the permutation group. This subgroup is the residual symmetry that remains after the breaking of replica symmetry. The Ansatz for the order parameter matrix corresponding to this residual symmetry is constructed as follows: The n × n (i.e. p0 × p0 ) matrix qαβ is divided into blocks of size p1 ×p1 , and a common value q0 is assigned to all matrix elements outside the diagonal blocks. Next the diagonal blocks are further divided into blocks of size p2 × p2 , the value q1 assigned to the matrix elements inside the diagonal blocks of size p1 × p1 but outside the diagonal blocks of size p2 × p2 , etc., down to the innermost blocks of size pR × pR , where the matrix elements are qR except along the very diagonal of the whole matrix where they are zero. Some formulae below (eqs. 32,34) become meaningless whenever the ratio of two subsequent p’s is 2 or 3. These cases would require a separate discussion which we can safely omit here, since in practical applications these cases will never appear. For our present purposes we can stipulate pk /pk+1 > 3, k = 0, 1, . . . , R. The solution of the stationary condition ∂q∂Fαβ = 0 is sought among the matrices which have the special form just described. This solution is a point in the 21 n(n − 1)-dimensional replica space, so it is, in fact, a vector. In the following, when we deal with genuine matrices acting on replica space, i.e. with quantities depending on two pairs of replica indices, we will actually call qαβ and similar quantities vectors. The association between the n × n symmetric matrix qαβ (with qαα = 0) and the column vector |qαβ i is evident: one lists the matrix elements above the diagonal of qαβ in any prescribed order (say, row by row) below each other. The representation of qαβ and other vectors of replica space by symmetric matrices remains, nevertheless, very useful, because it is much easier to display their special structure in the matrix form. Therefore, we shall use this matrix representation for all vectors appearing in this paper. For later reference we note here that the scalar product of two vectors, |rαβ i and |qαβ i, is, in matrix language, half the trace of the product of the corresponding matrices: 1 hrαβ |qαβ i = T r(rq). (1) 2 We now introduce the concept of the overlap between replica indices that will play a central role in the following: the overlap between α and β is k (notation: α ∩ β = k) if in the Parisi scheme qαβ = qk . The overlap α ∩ β defined this way can be regarded as a kind of hierarchical distance between replicas α and β, its values ranging from 0 (corresponding to the largest off-diagonal blocks of size p1 × p1 ) to R + 1 (corresponding to the diagonal, α = β). It is evident that any quantity f constructed of the q’s and depending on only two

replica indices (such as fαβ =

P γ

qαγ qγβ , for example) depends only on their overlap:

fαβ = f (α ∩ β). The metric generated by the overlaps is, by construction, ultrametric: whichever way we choose three replicas α, β, γ, either all three of their overlaps are the same (α ∩ β = α ∩ γ = β ∩ γ), or one (say α ∩ β) is larger than the other two, but then these are equal (α ∩ β > α ∩ γ = β ∩ γ). Furthermore, it also follows that any quantity f built of the q’s and depending on three replica indices, fαβγ , depends only on the overlaps α ∩ β, α ∩ γ, β ∩ γ, and since of these at most two can be different, fαβγ is, in fact, a function of only two variables, e.g. of α ∩ β and of the larger of the other two: fαβγ = f (α ∩ β; max{α ∩ γ, β ∩ γ}).

(2)

In the following we will also have to consider quantities depending on four replica indices, coming in two pairs: fαβ,γδ . A little reflection shows that such a quantity can always be parametrised as follows: α∩β,γ∩δ fαβ,γδ = fmax{α∩γ,α∩δ}, max{β∩γ,β∩δ} .

(3)

Admittedly, this parametrisation is less than perfect. Firstly, ultrametricity implies that of the six possible overlaps between α, β, γ and δ at most three can be different, which corresponds to the simple geometric fact that the edges of a tetrahedron having equilateral or isosceles faces can have at most three different lengths. Therefore, of the four variables on the r.h.s. of (3) at least two are the same. The resulting redundancy is the price we pay for the symmetry of the notation. Secondly, in all practical applications fαβ,γδ is symmetric w.r.t. exchanging α and β or γ and δ and also w.r.t. exchanging the two pairs: fαβ,γδ = fαβ,δγ = fβα,γδ = fγδ,αβ etc., and these symmetries are not manifestly reflected by the parametrisation (3). We prefer keeping the consequences of these symmetries in mind rather than overcomplicating the notation. The choice between the various types of solutions of the equation of state (identically zero, or constant qαβ , or replica symmetry broken qαβ ) is based on stability considerations. In order to decide the stability of a given solution, one has to diagonalize the Hessian or ∂2F (bare) self-energy matrix ∂qαβ ∂qγδ = Mαβ,γδ , evaluated at the stationary point. M is the prime example of a quantity depending on two pairs of replica indices, so it can be parametrised as shown in (3). M is obviously symmetric w.r.t. the exchange of the two pairs (αβ) and (γδ). Since qαβ = qβα and qαα = 0, M can be considered to depend on the ordered pairs α < β and γ < δ only, so it is a matrix of dimension 21 n(n − 1) × 12 n(n − 1). A symmetric matrix of this size has, in general, 41 n(n − 1)[ 21 n(n − 1) + 1] independent elements. This number is greatly reduced by ultrametricity. Below we list all the different kinds of matrix elements that can appear. When doing so, we will relax the ordering of the indices of Mαβ,γδ , and extend the definition to arbitrary combinations of the indices (except α = β and γ = δ) such as to make M symmetric w.r.t. exchanging α and β and/or γ and δ, Mαβ,γδ = Mαβ,δγ = Mβα,γδ = Mβα,δγ , in addition to the symmetry w.r.t. exchanging the pairs (αβ), (γδ). This extension is motivated by convenience: when

summations are to be performed on the indices of M the restrictions due to ordering can become very cumbersome. The matrix elements can be classified naturally in three categories: (i) Matrix elements of the first kind. These are the diagonal elements Mαβ,αβ , together with their variants Mαβ,βα , Mβα,αβ , etc. They depend on the overlap α ∩ β = i = 0, 1, 2, . . . , R only. Under the parametrisation (3) they are given by i,i Mαβ,αβ = MR+1,R+1

i = 0, 1, . . . , R.

(4)

There are, in general, R+1 different matrix elements in this category (instead of 21 n(n−1), the dimension of the matrix). (ii)Matrix elements of the second kind. These are off-diagonal, with one replica index in common between the two pairs. One example is Mαβ,αγ which, together with its exchanged variants (Mαβ,γα etc.), exhausts all possibilities. There are three cases: (a) α ∩ β = α ∩ γ = i ≤ β ∩ γ = j = 0, 1, . . . , R. Then i,i Mαβ,αγ = MR+1,j

,

j ≥ i.

(5)

Various exchanges of the replica indices either reproduce the same, or exchange the i,j lower variables R + 1 and j. (The parametrisation (3) is such that Mk,l is always symmetric in k and l.) (b) α ∩ β = β ∩ γ = i < α ∩ γ = j. i,j Mαβ,αγ = MR+1,i

,

j > i.

(6)

Exchanging replica indices in all possible ways reproduces either the same, or exchanges the lower variables, or exchanges i and j. Thus: j,i j,i i,j i,j = Mj,R+1 = MR+1,j = Mi,R+1 MR+1,i

,

i < j.

(7)

(c) α ∩ β = i > α ∩ γ = β ∩ γ = j. i,j Mαβ,αγ = MR+1,i

,

i > j.

According to (7), this is the same as (6) (rename i ↔ j). It is easy to see that the number of different matrix elements is at most (R + 1)2 in this class. (iii) Matrix elements of the third kind. These have four different replica indices, Mαβ,γδ . Considering all logically possible situations with α < β, γ < δ, α < γ, β < δ (corresponding to the matrix elements above the diagonal of M ), we find six possible cases altogether. (a) α ∩ β = i, γ ∩ δ = j, max{α ∩ γ, α ∩ δ} = max{β ∩ γ, β ∩ δ} = k with k ≤ min{i, j}. Then i,j j,i Mαβ,γδ = Mk,k = Mk,k , (8)

where the second equality follows from exchanging the two pairs (αβ) ↔ (γδ). (b) α ∩ β = i, γ ∩ δ = j, max{α ∩ γ, α ∩ δ} = max{β ∩ γ, β ∩ δ} = k, j < k ≤ i. Then i,j j,i Mαβ,γδ = Mk,k = Mk,j .

(9)

(c) α ∩ β = i, γ ∩ δ = j, max{α ∩ γ, α ∩ δ} = i, max{β ∩ γ, β ∩ δ} = k, j ≤ i < k. Then i,j j,i j,i Mαβ,γδ = Mi,k = Mk,j = Mj,k .

(10)

(d) α ∩ β = i, γ ∩ δ = j, max{α ∩ γ, α ∩ δ} = i, max{β ∩ γ, β ∩ δ} = k, i < j ≤ k. Then i,j j,i Mαβ,γδ = Mi,k = Mk,j .

(11)

(e) α ∩ β = i, γ ∩ δ = j, max{α ∩ γ, α ∩ δ} = i, max{β ∩ γ, β ∩ δ} = k, i < k < j.Then i,j j,i Mαβ,γδ = Mi,k = Mk,k .

(12)

(f) α ∩ β = γ ∩ δ = i, max{α ∩ γ, α ∩ δ} = k, max{β ∩ γ, β ∩ δ} = l, i < min{k, l}. Then i,i . Mαβ,γδ = Mk,l

(13)

In eqs. (8)-(13) the overlaps i,j,k,l can run through 0, 1, 2, . . . , R. Considering the cases (a)-(f) one can show that the number of different matrix elements of the third kind is (R + 1)3 . If all the 12 n(n−1) independent order parameter components were different, the matrix Mαβ,γδ would have of order n4 independent matrix elements. Parisi’s RSB scheme does not completely destroy the permutation symmetry of the replicas, however, it only reduces this symmetry to a particular subgroup of the group of permutations of n elements. It is this residual symmetry that is responsible for the tremendous reduction in the number of independent elements of the Hessian: instead of O(n4 ) we have, according to eqs. (4)-(13), only O(R3 ) different matrix elements, which, for large n, is exponentially small compared to n4 . The particular structure described above has been displayed in the example of the Hessian of the long-range spin glass. Matrices with an identical structure appear in many RSB theories. We shall call these matrices ultrametric matrices. From this point on we shall disregard the derivation and meaning of M , and will focus solely on its symmetries. It will be seen that these symmetries allow one to construct an irreducible representation for ultrametric matrices in that all those that have the same block sizes p0 , p1 , . . . , pR can be brought to a block diagonal form by the same similarity transformation and that no further reduction is possible without providing further information on the matrix elements. It will also be seen that the conditions ultrametricity imposes upon M are stringent enough to actually yield a large number of the eigenvalues in closed form. We shall also look into the problem of inversion of ultrametric matrices and shall find again that a large number of the components of the inverse can be obtained in closed form. Some of the results we compile here are not new, they were published by two of us some years ago in a very compact form [13]. In addition to rephrasing them and providing some background material we

also present a number of new results, especially with regards to the inversion problem. As long as n is an integer with the sequence of divisors p1 , p2 , . . . , pR as described, the matrix M is a well-defined mathematical object, and the problem of its diagonalisation belongs to the realm of standard mathematics. The present paper will be concerned with this well-posed problem. At a certain point one will, however, have to consider the replica limit n → 0, together with the analytic continuation in all the pi ’s and with the limit R → ∞, as proposed by Parisi [1]. These manipulations are at the present time of a purely formal character, certainly beyond the limits of well-established mathematics. After all these dubious steps one arrives at the problem of the diagonalisation of an integral operator with a set of particular symmetries. The results we get in the discrete case can all be easily transcribed on to this new, continuous problem. 3. The new basis In the previous section the components of an ultrametric matrix Mαβ,γδ have been given in the Cartesian coordinate system spanned by the basis vectors |µ, νi, (µ, ν) = (1, 2), (1, 3), . . . , (n − 1, n), which, similarly to the order parameter qαβ , can be represented by symmetric n × n matrices. Their matrix elements are  1, if µ = α, ν = β or µ = β, ν = α; Kr |µ, νiαβ = δ(µ,ν),(α,β) = (14) 0, otherwise. In order to bring M to a block diagonal form, we have to go over into a new basis. The new set of basis vectors can be inferred from the general structure of the eigenvectors described in [13] and, like those, can be naturally classified in three families. The first family The first family of basis vectors consists of R + 1 vectors labelled by i = 0, 1, . . . , R which, when represented by quadratic matrices, have identical nonzero elements on the ith level of the Parisi hierarchy and zeros everywhere else: − 21 X  1 n(pi − pi+1 ) |0; ii = |α, βi (15) 2 α∩β=i

where the |α, βi’s are the Cartesian unit vectors defined in (14). The meaning of the first label (0) will become clear shortly. The first family basis vectors form an (R + 1)-dimensional orthonormal set in replica space. The difference pi − pi+1 in the normalisation factor will appear so often in the following that it is worth giving it a name: pi − pi+1 = δi

,

i = 0, 1, . . . , R.

(16)

A straightforward but tedious calculation shows that the subspace spanned by the first family basis vectors is closed under the action of an ultrametric matrix. Therefore the linear combination R X f0 (i) |0; ii (17) |f i = i=0

is an eigenvector of M , provided the amplitudes f0 (i) are appropriately chosen. The conditions for these amplitudes (i.e. the eigenvalue equations) will be written up in the next section. Evidently, there will be R + 1 possible choices for the amplitudes, corresponding to R + 1 eigenvalues λm (0), m = 0, 1, . . . , R. In the case of a generic matrix M with no symmetries other than those dictated by ultrametricity, all these eigenvalues will be singlets, their multiplicity µ(0) = 1, and the eigenvectors orthogonal. In the following we shall often refer to the first family as the longitudinal or L family. The second family The second family will be broken down into several subfamilies, to be labelled by an index k = 1, 2, . . . , R + 1. The first family is, in several respects, nothing but the case corresponding to k = 0, which is why we used the label 0 in addition to i. The structure of the second family basis vectors is easiest to grasp graphically, so we define them in a series of figures. The vectors belonging to the k = 1 subfamily are shown in Figs. l,2. (Although, as we have already mentioned, the ratios of subsequent p’s must never be 2 or 3, in order to prevent the figures from occupying an excessive space, here and in almost all the figures to follow we have to illustrate the structure of eigenvectors by figures where some of these ratios are 3.) Consider the vectors shown in Fig. l. They have nonzero components only on the zeroth level of the Parisi hierarchy, but now not all these components are identical: they take two different values, A and B, arranged as shown in the figure. We shall denote these vectors as |1; 0; bi, where the first label is the value of k, the second is that level of the Parisi hierarchy where the vector has nonzero elements, and the third, b = 1, 2, . . . , n/p1 , shows which column and row of blocks is distinguished, i.e. which blocks have matrix elements B.

n

p1

B A

Fig. l: The vectors |k = 1, i = 0, bi. The label b = 1, 2, . . . , n/p1 shows which column and row of p1 × p1 blocks is distinguished. Identical shading means identical components. Blank means zero. Pn/p Now consider the sum of these vectors, b=11 |1; 0; bi. The distinction between the different blocks will obviously disappear in the sum, so it will be proportional to |0; 0i of the first family. However, we want to make each of the second family vectors orthogonal to the first family, so we have to choose the vector components, A and B so as to make the above sum vanish. If we choose, say, A = 1 then B must be   n 1 2− . (18) B= 2 p1

With this choice the |1; 0; bi vectors are all orthogonal to the first family and n/p1

X

|1; 0; bi = 0.

(19)

b=1

The evident symmetry between these vectors makes it obvious that they pairwise make the same angle which, together with (19), means that they span an ( pn1 − 1)-dimensional hypertetrahedron. It also follows that there is no further linear relationship between them, so if we discard one, say |1; 0; n/p1i, we will be left with

µ(1) = n



1 1 − p1 p0



(20)

linearly independent basis vectors. These will not be normalised, nor orthogonal, however. It would be an easy task to construct an orthonormal set out of them, but it would destroy their symmetry. We find it slightly more convenient to work with a biorthogonal set. For the same reason we need not worry about normalisation. It is an elementary exercise to show that the set 4p1 2 n (n − 2p1 )

  n |1; 0; bi − |1; 0; i , p1

b = 1, 2, . . . ,

n − 1, p1

(21)

is biorthogonal to the set |1; 0; bi. We now proceed, still within the k = 1 subfamily, to the next level of the Parisi hierarchy. The vectors |k = 1; i = 1; bi are shown in Fig. 2.

p2

p1

n

B A

Fig. 2.: The basis vectors |1; 1; bi.

Similarly to the previous case, orthogonality to the first family vector |0; 1i demands n/p1

X

|1; 1; bi = 0

b=1

which is satisfied if the vector components are chosen as A = 1,

B =1−

n . p1

(22)

Then the vectors |1; 1; bi span a µ(1)-dimensional hypertetrahedron again, with the associated biorthogonal set 2p1 2 n (p1 − p2 )

  n |1; 1; bi − |1; 1; i , p1

b = 1, 2, . . . ,

n − 1. p1

(23)

The construction proceeds along similar lines: filling in the ith level of the Parisi hierarchy we find the same µ(1)-dimensional tetrahedra with the same orthogonality conditions (22) applying for any i ≥ 1 (only the case i = 0, eq. (18), is different). This way we will have, altogether, µ(1) (R + 1) independent basis vectors |1; i; bi,

i = 0, 1, . . . , R and

b = 1, 2, . . . ,

n − 1, p1

making up the k = 1 subfamily. They are each orthogonal to the first family, two of them belonging to different values of i are also orthogonal, but two such vectors with the same i and different b’s are not. It can now be shown again that the subspace spanned by the R + 1 vectors |1; i; bi, i = 0, 1, . . . , R, for b fixed is an invariant subspace of an arbitrary ultrametric matrix. Therefore the linear combination

|f i =

R X

f1 (i) |1; i; bi

i=0

with appropriately chosen amplitudes f1 (i), independent of b, will be an eigenvector. There will be R+1 choices for these amplitudes, yielding R+1 eigenvalues λ1 (m), m = 0, 1, . . . , R, in the k = 1 subfamily. Each of these will be µ(1)-fold degenerate, according to the free choice of b.

We now turn to the k = 2 subfamily. Some of the k = 2 type vectors with the i = 0 Parisi level filled are shown in Fig. 3. p

1

p

2

n

B A

etc.

etc.

Fig. 3: Some vectors of the k = 2 subfamily (i = 0).

These vectors will be labelled as |2; 0; a, bi where a = 1, 2, . . . , n/p1 shows which column of p1 × p1 sized blocks has nonzero elements and b = 1, 2, . . . , p1 /p2 shows which column Pp1 /p2 |2; 0; a, bi must of p2 × p2 sized blocks is distinguished inside column a. The sum b=1 vanish again for any fixed a, otherwise it would be a linear combination of k = 0 and k = 1 subfamily type vectors. This orthogonality condition demands that we choose A = 1,

B =1−

p1 p2

(24)

leaving p1 /p2 − 1 independent vectors (spanning a tetrahedron again) for any fixed a. It is easy to see that with this choice the vectors |2; 0; a, bi will not only be orthogonal to each of the previous families (with k = 0, 1) but they will also be orthogonal to the vectors |2; 0; a′, b′ i with a 6= a′ and any b′ . Some k = 2, i = 1 vectors are shown in Fig. 4.

p

2

p1

n

BA

etc.

etc.

Fig. 4.: Some vectors of the k = 2 subfamily (i = 1). The orthogonality conditions now read A = 1,

1 B= 2

  p1 2− . p2

We can go on to build |2; i; a, bi, i = 0, 1, . . . , R in a similar manner. The subspace of these vectors for fixed a and b will be closed under the action of an ultrametric matrix M . Thus R X

f2 (i) |2; i; a, bi

i=0

will be an eigenvector with R + 1 choices for the amplitudes and with eigenvalues λ2 (m), m = 0, 1, . . . , R. We have seen that for a given a we have pp12 − 1 linearly independent choices for the basis  vectors,  while for different a’s they are already orthogonal. That p1 n means we have p1 p2 − 1 , i.e. µ(2) = n



1 1 − p2 p1



independent basis vectors for any i. The total dimension of the k = 2 subfamily is thus µ(2) (R + 1), and the multiplicity of the k = 2 eigenvalues is µ(2). The generalisation is now obvious. In the k th subfamily (k = 1, 2, . . . , R + 1) we have vectors labelled by four indices: |k; i; a, bi. This vector has nonzero elements only on the

ith level of the Parisi hierarchy, and there only inside one column and row of blocks of size pk−1 × pk−1 . There are n/pk−1 such columns, and the label a = 1, 2, . . . , n/pk−1 shows which is the one in question. The vector components inside these blocks are all 1’s, except inside a distinguished column and row of blocks of size pk × pk where they are *   1 pk−1 B= 2− , if i = k − 1 2 pk (25) pk−1 B =1− , otherwise. pk p

The distinguished columns of pk ×pk blocks are labelled by the last index, b = 1, 2, . . . , k−1 pk . With the choice (25) the vectors |k; i; a, bi are orthogonal to all the previous subfamilies with k−1, k−2, . . . etc. down to the first family. Within the k th subfamily vectors belonging to different i’s and a’s are also orthogonal, while those with a fixed k, i, a and different b’s make a ( pk−1 − 1)-dimensional hypertetrahedron. The number of linearly independent pk   pk−1 n vectors for a given k and i will then be pk−1 pk − 1 , i.e. µ(k) = n



1 1 − pk pk−1



.

(26)

The biorthogonal set associated with the tetrahedral groups of vectors belonging to a given triplet k, i, a is   1 pk−1 pk g i (27) |k; i; a, bi − |k; i; a, |k; i; a, bi = 2 pk−1 g (k) pk i (k)

where the weight gi

is defined as    pi − pi+1 , (k) gi = 41 (pk−1 − 2pk ),   1 (p − p ), 2

i

i+1

i k − 1.

For fixed k, a, b the R + 1 vectors |k; i; a, bi, i = 0, 1, . . . , R, form an invariant subspace of any ultrametric matrix, so we will have eigenvectors of the form R X

fk (i) |k; i; a, bi

(29)

i=0

with amplitudes fk (i) independent of a,b. The corresponding eigenvalue equations will give R+1 possible values for fk (i), and the eigenvalues λm (k), m = 0, 1, . . . , R, will be µ(k)-fold degenerate. Sometimes the second family is also called the anomalous or A family. * We take the opportunity to   correct a misprint in eq. 4 in [13] here. The B for i = k − 1 p there instead of the correct expression in (25). was given as B = 21 1 − k−1 pk

So far from the 21 n(n − 1)-dimensional replica space we have split off the (R + 1)dimensional invariant subspace of the first family, the (R + 1)-dimensional subspaces, µ(1) in number, of the k = 1 subfamily, etc., up to k = R + 1, that is we have decomposed our linear space into R+1 X

µ(k) = n

k=0

(R + 1)-dimensional invariant subspaces plus the vast space, of dimension n(R + 1), orthogonal to the first and second families.

1 n(n 2

− 1) −

The third family The third family, often called the replicon or R family, comprises everything remaining after splitting off the first two families. It is a most remarkable fact, and a direct consequence of the stringent conditions ultrametricity imposes upon a matrix, that the third family can be decomposed into invariant subspaces of dimension 1, i.e. directly into eigenvectors. This also means that the third family eigenvalues that, for large n, represent the overwhelming majority of all the eigenvalues can be obtained in closed form, in terms of the matrix elements, for any ultrametric matrix. The third family eigenvectors were given in a concise form in [13]. We provide a little more detail here which will become important when we invert the matrix M . The third family consists of several subfamilies labelled by three integers r = 0, 1, . . . , R k, l = r + 1, r + 2, . . . , R + 1. There will be several degenerate vectors in each subfamily. They will be labelled by three, five, or seven more indices, as the need arises. A common property of all third family vectors is that they have nonzero components only inside one single diagonal block of size pr × pr , which also gives the significance of the label r above. The labels k and l specify further structural details that are best displayed on a series of figures again. In the following we shall exhibit only that pr × pr -sized block over which the vector components are not all zero. The r, k = r + 1, l = r + 1 subfamily The structure of the nonvanishing block is shown in Fig. 5. These vectors take three further labels to specify them |r; r + 1, r + 1; a, b, ci. The index a = 1, 2, . . . , n/pr shows which of the n/pr diagonal pr × pr blocks has nonvanishing elements. Inside this block all the components belonging to the diagonal pr+1 × pr+1 blocks vanish again. Of the off-diagonal pr+1 × pr+1 blocks those belonging to two columns and rows are distinguished and a further distinction is made between the blocks at the crossing of a distinguished column and row and the rest. The indices b, c = 1, 2, . . . , pr /pr+1 , b 6= c, label the two distinguished columns. In all, we then have three different vector components in this subfamily, as shown in the figure.

pr+1

A

B

pr

C

Fig. 5.: The third family vector |r; r + 1, r + 1; a, b, ci. Identical shading means identical components, blank means zero. b = 1, c = 3 in the figure. Orthogonality to the previous families requires that pr /pr+1

X

|r; r + 1, r + 1; a, b, ci = 0

b=1 b6=c

(31) pr /pr+1

X

|r; r + 1, r + 1; a, b, ci = 0.

c=1 c6=b

It follows from (31) that the sum of vector components in each row should vanish, giving us two equations for the three numbers A, B, C, the third being determined by normalization. They work out to be: pr+1 A B=− pr − 2pr+1 C= A2 =

2p2r+1 A (pr − 2pr+1 )(pr − 3pr+1 ) pr − 3pr+1 2 pr+1 (pr − pr+1 )

(32)

.

With this we have determined the eigenvectors with r, r + 1, r + 1 completely. The corresponding eigenvalues will be written up in the next section. For a given position of the   pr 1 pr pr × pr block, i.e. for a given a, the orthogonality conditions (31) leave 2 pr+1 pr+1 − 3 vectors linearly independent. This number, multiplied by pnr , the number of choices for a, gives the multiplicity of this class: µ(r; r + 1, r + 1) =

1 pr − 3pr+1 n , 2 p2r+1

r = 0, 1, . . . , R.

(33)

Eigenvectors with r, k > r + 1, An example is shown in Fig. 6.

l =r+1 A

C }p

k

B

pr+1

}

pk-1

D

pr

Fig. 6.: An eigenvector of the class |r; k ≥ r + 2, l = r + 1; a, b, c, d, ei. Such a vector is constructed as follows. One chooses a diagonal block of size pr × pr , labelled by a = 1, 2, . . . , n/pr , as before. Inside this block one chooses two columns and rows of blocks of size pr+1 × pr+1 , say the bth and the cth , such that c > b (b = 1, c = 2 in Fig. 6). Inside the blocks in the bth column and row one now chooses a strip of blocks of size pk−1 × pk−1 , say the dth , as shown in the figure. All the vector components outside this strip will be zero. Inside the strip one chooses a strip, the eth , of blocks of size pk × pk . Finally the vector components A, B, C, D are arranged, as shown, according to whether they belong to the strip of width pk or are outside, and also whether they belong to the cth blocks or not. Orthogonality to previous families again requires, as throughout the third family, that the sum of components in each row vanish. This immediately gives pk A pk−1 − pk pr+1 C=− A pr − 2pr+1 pk pr+1 D= A (pk−1 − pk )(pr − 2pr+1 )   1 1 pr − 2pr+1 2 − A = pr+1 (pr − pr+1 ) pk pk−1 B=−

(34)

for the components of a normalised vector of this class. Considering the various choices for the parameters a, b, c, d, e one finds that the total number of linearly independent vectors of this kind is   1 pr − 2pr+1 1 1 µ(r; k, r + 1) = n − , k = r + 2, r + 3, . . . , R + 1. (35) 2 pr+1 pk pk−1

The subfamily r; k = r + 1, l > r + 1 is obtained by the same construction but with c < b; this subfamily will be similar to the r; k > r + 1, l = r + 1 subfamily in every respect, but it will be orthogonal to it. Eigenvectors with r, k > r + 1, l > r + 1 The construction of these vectors is shown in Fig. 7. It begins again by choosing one diagonal block of size pr × pr , labelled by a. This can be done in n/pr different ways. Next, inside this block one chooses two symmetrically positioned off diagonal blocks of size pr+1 × pr+1 . This takes b, c = 1, 2, . . . , pr /pr+1 , and the number of inde two indices: 

pr pr pendent choices is 12 pr+1 pr+1 − 1 , because of the symmetry of the matrix representing the eigenvector. Now the off diagonal pr+1 × pr+1 block above the diagonal is cut into rectangles of horizontal size pl−1 and vertical size pk−1 and the one below the diagonal is cut similarly with the horizontal and vertical dimensions exchanged.

c

p pr

l-1

}

pr+1

pr+1 }pk-1

A

C pk

b p

pk-1 pl-1

l

B (a)

(b)

D (c)

Fig. 7.: The construction of an eigenvector with r, k > r + 1, l > r + 1. The structure of the vector is shown in three stages of finer and finer resolution. The block with indices c, b in (a) is obtained by reflection to the diagonal from block b, c. One of these rectangles is chosen, which again takes two labels: d = 1, 2, . . . , pr+1 /pl−1 , and e = 1, 2, . . . , pr+1 /pk−1 , and can be done in p2r+1 /(pl−1 pk−1 ) ways. We have nonzero components only inside these rectangles. Their structure is shown in Fig. 7(c), and can evidently be characterised by two further indices f = 1, 2, . . . , pl−1 /pl , and g = 1, 2, . . . , pk−1 /pk . These vectors are orthogonal to each other in all the indices except for f and g. For fixed f the set with different g’s forms the usual tetrahedron again, and −1)×( pl−1 −1) independent the same is true for fixed g in the f ’s, so we are left with ( pk−1 pk pl choices for f and g. All these taken together give a multiplicity    1 1 1 1 1 , k, l = r + 2, r + 3, . . . , R + 1, − − µ(r; k, l) = n (pr − pr+1 ) 2 pk pk−1 pl pl−1 (36) while the usual orthogonality conditions (the sum of vector components in each row and each column vanishes) give us the following values for the components of a normalised

vector of the r, k > r + 1, l > r + 1 type: pk A pk−1 − pk pl C=− A pl−1 − pl pk pl A D= (pk−1 − pk )(pl−1 − pl )    1 1 1 1 2 A = − − . pk pk−1 pl pl−1 B=−

(37)

With this we have given a full description of the third family eigenvectors. To show that these forms are indeed reproduced under the action of an ultrametric matrix takes, of course, a lot of algebra. It is impossible for us to go into details on this, but we think a little experimentation in a simple special case like R = 1 will convince the reader that the proof is quite straightforward though certainly not very short. As we have seen, there is a high degree of degeneracy within each subfamily (r; k, l). These subfamilies are all orthogonal to each other (and also to the first and second families, of course), and some of the degenerate vectors within a given (r; k, l) subfamily are also orthogonal, but some form the by now familiar tetrahedral sets in more than one index. It would not be difficult to orthogonalise these vectors, or, alternatively, to construct biorthogonal sets to them. We refrain from doing both: the loss in symmetry would be considerable and the gain virtually nothing. The only occasion when we might need a properly orthonormalised set is when later we construct the ”replicon” components of the inverse of M from the spectral resolution. It will turn out, however, that the orthogonalisation can be circumvented even there and the ultrametric symmetries of M (and of its inverse) will allow us to deduce the full contribution of the whole (r; k, l) subfamily to the inverse from the knowledge of a single vector belonging to that subfamily. This vector will be called the representative vector of the subfamily. We can choose any of the µ(r; k, l) degenerate vectors to be the representative vector. Suppose we have made our choice. Some of the vectors in the (r; k, l) subfamily will be orthogonal to the selected vector from the beginning. These will, as a rule, have zero components where the selected vector has nonzero ones, in particular, they will have zeros where the components we called A’s in the description of the third family vectors, i.e. those in the darkest shaded areas in the figures, are to be found in the selected vector. Now consider the vectors which are not orthogonal to the selected one. In order to orthogonalise them to the representative vector we form some linear combinations. Our key observation is now that, due to the defining properties of the third family vectors (namely, that the sums of vector components in each row are zero), any linear combination that is orthogonal to the representative vector will have zero components where the representative vector has its components A. The proof of this will also be left as an exercise to the reader. The last issue to be settled in this section is the total multiplicity of our basis vectors. In the first two families we found n (R + 1) independent basis vectors. In the subfamilies

(r; k, l), for fixed r, there are a total of R+1 X

µ(r; k, l) =

k,l=r+1

1 n(pr − pr+1 − 2) 2

(38)

independent vectors. This summed over r gives the total number of basis vectors in the third family: R X 1 1 n(pr − pr+1 − 2) = n(n − 2R − 3). (39) 2 2 r=0

Added to n (R + 1) this gives 12 n(n − 1), the dimension of replica space. Our set of basis vectors is therefore complete. 4. The block diagonal form of ultrametric matrices Having constructed a complete set of basis vectors we can build a matrix S with columns made of these vectors and transform M to this new basis by ˜ = S −1 M S. M

(40)

S is not an orthogonal transformation (because the basis vectors are not all orthogonal), ˜ will not be symmetric. The rows of the inverse S −1 in the first family sector will so M be made of the bra vectors corresponding to the first-family-like basis vectors |0; ii, in the second family sector they will be the biorthogonal vectors given in (27). In the third family sector we do not really need to construct the matrix S −1 at all, since there the basis vectors are the eigenvectors themselves. Since the various families and subfamilies were constructed in such a way that they are invariant subspaces of M , the transformed ˜ will have a block diagonal form: along the diagonal we will have a string of n matrix M (R +1)×(R +1) matrices, the first corresponding to the first family, the next µ(1) identical matrices corresponding to the k = 1 subfamily in the second family, etc. through µ(k) identical blocks for the k th subfamily up to k = R + 1. This string of matrices will be followed by the string of the third family eigenvalues coming in groups of µ(r; k, l) identical numbers corresponding to the subfamilies (r; k, l). The third family or replicon eigenvalues are obtained as a byproduct of checking that the third family vectors given in the previous section are eigenvectors indeed. In the (r; k, l) subfamily one obtains the closed expression λ(r; k, l) =

R+1 X s=k

ps

R+1 X

r,r r,r r,r r,r pt (Mt,s − Mt−1,s − Mt,s−1 + Mt−1,s−1 )

t=l

(41)

r = 0, 1, . . . , R

k, l = r + 1, r + 2, . . . , R + 1, giving the replicon eigenvalues directly in terms of the matrix elements and of the p’s characterising the structure of M . Eq. (41) has been written up already in [13].

˜ will be labelled by k = 0, 1, . . . , R + 1 The (R + 1) × (R + 1) diagonal blocks of M (k) as M , k = 0 corresponding to the first family, k > 0 to the subfamilies in the second family. The matrix elements of M (0) can be obtained by sandwiching M between two first family vectors (0) Mr,s = h0; r|M |0; si, (42) those of M (k) by sandwiching M between a second family vector and one from the biorthogonal set given in (27): (k) Mr,s = hk; g r; a, b|M |k; s; a, bi (43) and can really be obtained again as byproducts when verifying the invariance of the various subfamilies under the action of the matrix M . We can therefore simply state the result: (k) Kr Mr,s = Λ(k, r)δr,s + gs(k)

   (k) (k) 1 1 Kr Kr ∆r + 2 pk−1 δr,k−1 ∆s + 2 pk−1 δs,k−1 δr δs

Kk (r, s).

(44)

(k)

Of the symbols appearing here δr and gs have already been defined in (16) and (28), (k) Kr respectively. δr,s is the Kronecker symbol, while ∆r is

∆(k) r

 1   2 δr , = 12 (δk−1 − pk ),   δr ,

r k − 1.

(45)

The diagonal part Λ(k, r) is related to the third family eigenvalues: Λ(k, r) =



λ(r; r + 1, k), k >r+1 λ(r; r + 1, r + 1), k ≤ r + 1.

(46)

Now we define the kernel Kk (r, s) appearing in (44): r s 1 X 1X 1 r,s r,s r,s r,s ) − Mr,j−1 pj (Mj,j − Mj−1,j−1 ) + pj (Mr,j Kk (r, s) = 4 4 2 j=r+1 j=k

+

R+1 X

r,s r,s ), − Mr,j−1 pj (Mr,j

(47)

k − 1 ≤ r ≤ s,

j=s+1 r s 1 1 X 1X s,r s,r s,r s,r pj (Mj,j − Mj−1,j−1 ) + pj (Mj,j − Mj−1,j−1 ) Kk (s, r) = 4 4 2 j=r+1 j=k

+

R+1 X

j=s+1

s,r s,r ), − Ms,j−1 pj (Ms,j

k − 1 ≤ r ≤ s,

(48)

R+1 s X 1 1X r,s r,s r,s r,s ), − Mr,j−1 pj (Mr,j Kk (r, s) = pj (Mr,j − Mr,j−1 ) + 4 2

r ≤ k − 1 ≤ s, (49)

j=s+1

j=k

s R+1 X 1 1X s,r s,r s,r s,r pj (Mj,j − Mj−1,j−1 ) + pj (Ms,j − Ms,j−1 ), Kk (s, r) = 4 2 j=s+1

r ≤ k − 1 ≤ s, (50)

j=k

R+1 X 1 r,s r,s ), − Mr,j−1 Kk (r, s) = pj (Mr,j 4

r ≤ s ≤ k − 1,

(51)

r ≤ s ≤ k − 1.

(52)

j=k

R+1 X 1 s,r s,r ), − Ms,j−1 Kk (s, r) = pj (Ms,j 4 j=k

(Possible empty sums here and in the following are understood to be zero. For k = 0, r,s terms with M−1,−1 may occur in the above formulae. They are, by definition, zero too.) (k)

With this the matrix elements Mr,s in the new representation have been expressed in terms of the matrix elements in the original representation. The problem of finding the eigenvalues of M in the first two families has thus been broken down into the problem of finding the spectrum of each of the M (k) ’s. As we have already noted, due to the nonorthogonality of the transformation S, the M (k) ’s are not symmetric. Using the symmetries between the various components of M as described in Sec. 2, one can show, however, that, although we have given the expressions for both Kk (r, s) and Kk (s, r) in the various cases for completeness, the kernel Kk is, in fact, symmetric. Therefore, the (s) asymmetry of M (k) is carried solely by the factor gk in eq. (44), and we can, if we wish, reduce the eigenvalue problem of M (k) to that of the manifestly symmetric matrix 1

gr(k) 2

(k)

Mr,s (k)

gs

1

gs(k) 2 .

Let us now spell out the eigenvalue equation of M (k) : R X

(k) Mr,s fk (s) = λ(k)fk (r)

s=0

reads in the two cases r < k − 1 and r ≥ k − 1, respectively, as R 1 X 1 Λ(k, r)fk (r) + Kk (r, s)fk (s)δs + (δk−1 − pk )Kk (r, k − 1)fk (k − 1) 4 8 s=0

s6=k−1

= λ(k)fk (r),

r < k−1

R 1 X 1 Λ(k, r)fk (r) + Kk (r, s)fk (s)δs + (δk−1 − pk )Kk (r, k − 1)fk (k − 1) 2 4 s=0

s6=k−1

= λ(k)fk (r),

r ≥ k − 1.

(53)

This is a set of R + 1 homogeneous linear equations for any given k = 0, 1, . . . , R + 1. (In the first family, i.e. for k = 0, only the equation with r > k − 1 applies, with the third term on the l.h.s. discarded.) The solutions for the fk ’s give the amplitudes mentioned in the preceding section, while the λ(k)’s are the corresponding eigenvalues. It is evident from the definition that the space of ultrametric matrices (belonging to the same series p0 , p1 , . . . , pR ) is closed under addition. That it is also closed under multiplication is easiest to see from the existence of the common similarity transformation S that brings any two such matrices to blockdiagonal form simultaneously. Suppose we are given two ultrametric matrices M and M ′ , with the associated kernels ′ Kk and Kk′ and replicon eigenvalues λR and λR from which we have the corresponding Λ’s as given in (46). Then in the block diagonal representation of the product M M ′ we will find for the diagonal blocks in the LA sector:    (k) (k) 1 1 Kr Kr R ∆r + 2 pk−1 δr,k−1 ∆s + 2 pk−1 δs,k−1 X (k) (k) ′ Mr,t Mt,s = gs(k) × δr δs t=0 ) ( R 1 X (k) (54) ∆t Kk (r, t)Kk′ (t, s) + Λ(k, r)Kk′ (r, s) + Λ′ (k, s)Kk (r, s) 2 t=0 Kr + δr,s Λ(k, r)Λ′ (k, r)

where we have used that by (16), (28), (45)

(k)

gt



(k)

∆t

Kr + 21 pk−1 δt,k−1

δt2

2



1 (k) ∆ , 2 t

(55)

whereas in the R sector we will evidently find the product of the replicon eigenvalues: λ(r; k, l)λ′(r; k, l). In particular, if M ′ is the inverse of M , we have

R P

t=0

1, so from (54) we get

(k)

(56) ′(k)

Kr Mr,t Mt,s = δr,s and Λ(k, r)Λ′ (k, r) =

R

1 X (k) ∆ Kk (r, t)Kk′ (t, s) + Λ(k, r)Kk′ (r, s) + Λ′ (k, s)Kk (r, s) = 0. 2 t=0 t

(57)

For a given M , i.e. for a given Kk and Λ(k, r), this is an equation for the kernel Kk′ of the inverse matrix. In the general case (57) is still a matrix equation which is typically difficult to solve, though (57) certainly has the merit of reducing the problem of the inversion of a 12 n(n − 1) × 21 n(n − 1) dimensional matrix to that of inverting R + 2 much smaller matrices (corresponding k = 0, 1, . . . , R + 1 in (57)) of size (R + 1) × (R + 1). In some important special cases, however, further progress can be made and both the solution of

the eigenvalue equations (53) and the inversion of the matrix M can be carried through to the end [12]. Coming back to eq. (57) let us assume now that we have somehow succeeded in solving it for Kk′ . With this the inversion of M is, however, not yet completed, because normally we need the inverse in the original Cartesian coordinate system, i.e. we need the components of M ′ given the kernel Kk′ . This means we have to invert the formulae (47)-(52), or, to put it even more simply, we have to turn (40) around like ˜ S −1 . M = SM

(58)

In the LA sector this is a standard operation: we have all the basis vectors and their biorthogonal counterparts so that we explicitly know the corresponding blocks in S and S −1 . This is not the case in the R sector where we have neither orthogonalised nor ˜ , this did biorthogonalised our basis vectors. In the direct transformation, from M to M not cause a problem, because the third family basis vectors being eigenvectors we knew ˜ would be the eigenvalues themselves. In in advance that the corresponding ”blocks” of M ˜ to M , however, we would definitely need the missing the inverse transformation, from M −1 blocks in S in order to determine the contribution of the replicon family to the various components of M . It is at this point that the concept of the representative vector introduced in the previous section becomes important. We do not think we should dwell upon how the blocks of the three matrices in (58) have to be multiplied in the sector where they are known. We have to explain, however, how the replicon contribution to (58) can be obtained from the representative vectors without actually knowing the corresponding block in S −1 . The Appendix is devoted to this problem. In what follows we will state our results for the nine different types of matrix elements ˜ . In of M discussed in Sec. 2, in terms of the matrix elements of the blockdiagonal form M each case we shall give the result in two different forms: first as a sum of two terms, one coming from the LA sector, the other from the replicon, and second, in a form where some most remarkable cancellations between these two have been effected. In the discrete case, where n, R, and all the pk ’s are integers, these cancellations may seem coincidental. We note, however, that in the continuous limit they acquire a fundamental importance [13]. In order to display these cancellations we partition the third family multiplicities as follows: µ(r; k, l) ≡ µreg (r; k, l) + µsing (r; k, l) (59) where 1 pr µreg (r; r + 1, r + 1) = n 2 pr+1 µsing (r; r + 1, r + 1) = −

r+1 X



1 pr+1

1 − pr



,

(60)

µ(k),

(61)

k=0

1 pr − pr+1 µreg (r; r + 1, k) = µreg (r; k, r + 1) = n 2 pr+1



1 1 − pk pk−1

k > r + 1,



, (62)

1 µsing (r; r + 1, k) = µsing (r; k, r + 1) = − µ(k), 2

k >r+1

(63)

and µreg (r; k, l) for k, l > r + 1 is the full µ(r; k, l) itself, as given in (36), so µsing (r; k, l) = 0,

k, l > r + 1.

(64)

In (61) and (63) µ(k) is the second family multiplicity. In the discrete case the subscripts ”regular” and ”singular” have no particular significance; in the continuous limit, however, µsing will be associated with terms that become meaningless but disappear from the theory due to the cancellations mentioned above. We now list the results: r,r = MR+1,R+1

R+1 X k=0

R+1 X 2µ(k) (k) Mr,r + nδr

R+1 X

k=r+1 l=r+1

2µ(r; k, l) λ(r; k, l) nδr

(65a)

where the first term is the LA, the second the R contribution as announced. Substituting (44) for M (k) and splitting µ as µ = µreg + µsing we see that the Λ(k, r) term coming from the LA cancels the µsing contributions from the R family exactly. So we have the alternative form: r,r MR+1,R+1

=

R+1 X k=0

R+1 X µ(k) (k) ∆r Kk (r, r) + nδr

R+1 X

k=r+1 l=r+1

2µreg (r; k, l) λ(r; k, l). nδr

(65b)

!

(66a)

Similarly: r,r MR+1,t

=

R+1 X k=0

+

(k)

2µ(k) ∆t (k) Mr,r nδr δt

R+1 X

R+1 X

k=r+1 l=r+1

=

R+1 X k=0

+

2µ(r; k, l) λ(r; k, l) nδr (k)

(k)

(l)

∆ ∆t + t −1 δt δt

(k)

µ(k) ∆r ∆t Kk (r, r) n δr δt

R+1 X

R+1 X

k=r+1 l=r+1

2µreg (r; k, l) λ(r; k, l) nδr

! (l) (k) ∆t ∆t + −1 , δt δt

t > r.

In the special case t = r of the above, things work out slightly differently: r,r MR+1,r =

=

R+1 X k=0

(k)

1 Kr 2µ(k) ∆r − 4 pk−1 δr,k−1 (k) Mr,r Kr nδr δr − 12 pk−1 δr,k−1

(66b)

n 2µ(r; r + 1, r + 1) pr+1 λ(r; r + 1, r + 1) − pr − 2pr+1 nδr R+1 o X 2µ(r; r + 1, k) λ(r; r + 1, k) + nδr

(67a)

k=r+2

=

R+1 X k=0

(k)

1 Kr µ(k) ∆r − 4 pk−1 δr,k−1 ∆(k) r Kk (r, r), Kr nδr δr − 12 pk−1 δr,k−1

(67b)

so that, as we see, the Λ term coming from the LA now cancels the whole replicon contribution, not just the one with µsing . The next item to be considered is a component of M with different upper indices, so that there is no replicon contribution to it:    1 1 Kr Kr R+1 (k) δ − p δ p δ δ − X s 2 k−1 s,k−1 Mr,s µ(k) r 2 k−1 r,k−1 r,s = MR+1,r (k) n δr δs gs k=0

=

R+1 X k=0

(k)

(k)

µ(k) ∆r ∆s Kk (r, s), n δr δs

s 6= r.

Now we turn to the matrix elements of the third kind: ! R+1 (k) X 2µ(k) ∆t r,s (k) Mt,t = 2 − 1 Mr,s nδs δt k=0 ! R+1 (k) X µ(k) ∆t Kk (r, s) 2 −1 , = n δt

(68)

t < r, s.

(69)

k=0

Note that this holds even for r = s, because R+1 X k=0

(k)

∆ µ(k) 2 t − 1 δt

!

= 0,

therefore the diagonal part Λ in M (k) never contributes. Neither does the replicon, because the upper indices are larger than the lower ones in here. The next four items are off-diagonal in the upper indices, so they do not receive contributions from the replicon family.   ! 1 Kr R+1 (k) (k) p δ δ − X k−1 s,k−1 s 2 µ(k) ∆ Mr,s r,s Mt,t = 2 t −1 (k) n δs δt gs k=0 ! R+1 (k) X µ(k) ∆(k) ∆t s 2 s < t < r, (70) − 1 Kk (r, s), = n δs δt k=0

r,s Mr,t

=

R+1 X k=0

=

R+1 X k=0

   1 1 Kr Kr δ − p δ p δ δ − s k−1 k−1 r r,k−1 s,k−1 2 2 µ(k) n

(k)

(k)

(k)

!

R+1 X

µ(k) n

! (k) ∆r 2 − 1 Kk (r, s), δr

R+1 X

1 Kr µ(k) δr − 2 pk−1 δr,k−1 n δr

=

k=0

k=0

R+1 X k=0

(k)

µ(k) ∆r n δr

!

(k)

Mr,s (k)

gs

r < s < t,

(71)

(k)

∆r 2 −1 δr

k=0

=

(k)

µ(k) n

r,s Mr,r =

r,s Mr,s =

δr δs !

∆ 2 t − 1 Kk (r, s), δt

µ(k) ∆r ∆s n δr δs R+1 X

(k)

∆ 2 t −1 δt

Mr,s (k)

gs

(72) r < s,

(k)

∆s 2 −1 δs

! (k) ∆s 2 − 1 Kk (r, s), δs

!

(k)

Mr,s (k)

gs

(73) r < s.

In the next two cases again a complete cancellation takes place between the Λ term in the LA and the replicon: r,r Mr,r

=

R+1 X k=0

(k)

1 Kr 2µ(k) ∆r + 2 pk−1 δr,k−1 Kr nδr δr − 21 pk−1 δr,k−1

! (k) ∆r (k) 2 − 1 Mr,r δr

2p2r+1 2µ(r; r + 1, r + 1) λ(r; r + 1, r + 1) nδr (pr − 2pr+1 )(pr − 3pr+1 ) ! R+1 (k) X µ(k) ∆r = 2 − 1 Kk (r, r), n δr

+

(74a) (74b)

k=0

R+1 X

(k)

1 (k) (k) Kr 2µ(k) ∆r − 4 pk−1 δr,k−1 ∆r ∆s r,r Mr,s = + −1 Kr nδr δr − 12 pk−1 δr,k−1 δr δs k=0 n 2µ(r; r + 1, r + 1) pr+1 − λ(r; r + 1, r + 1) pr − 2pr+1 nδr

!

δr (k) ∆r

(k) Mr,r

∆r δs + ∆s δr − δr δs o 2µ(r; r + 1, k) λ(r; r + 1, k) (k) nδr ∆r δs k=r+2 ! R+1 1 (k) (k) Kr X µ(k) ∆(k) r − 4 pk−1 δr,k−1 ∆r ∆s + − 1 Kk (r, r), = Kr n δr δs δr − 12 pk−1 δr,k−1

+

R+1 X

k=0

(k)

(k)

(75a)

s > r. (75b)

Finally, in the last type of component the Λ term from the LA cancels the µsing term in the R: ! R+1 (k) (k) X 2µ(k) ∆ ∆ s r,r (k) = Ms,t + t − 1 Mr,r nδr δs δt k=0 ! ! R+1 (k) (l) X R+1 X 2µ(r; k, l) ∆s ∆t 2 −1 2 − 1 λ(r; k, l) (76a) + nδr δs δt k=r+1 l=r+1 ! R+1 (k) (k) X µ(k) ∆ ∆ s = ∆(k) + t − 1 Kk (r, r) nδr r δs δt k=0 ! ! R+1 (k) (l) X R+1 X 2µreg (r; k, l) ∆s ∆t + 2 −1 2 − 1 λ(r; k, l) s, t > r. (76b) nδr δs δt k=r+1 l=r+1

As we have mentioned, in the equations from (65) to (76) the formulae denoted (a) give the true partition of the contributions between the LA and R families. If anyone tried to reproduce these results, they would inevitably get them in this form, and we give them here partly as signposts. In most applications the origin of the terms is completely immaterial, however, so that when using these formulae, one will clearly apply the (b) forms, where the cancellations have been performed. The names one attaches to these terms are also largely a matter of convention: in the papers [14] and [15] where analogous formulae were derived for the propagators two of us used the name LA for the first terms and the name R for the second terms in the (b) forms. We also see that there is nothing mysterious about the cancellations: the diagonal matrix elements of M (k) contain the replicon eigenvalue and this piece partially or completely cancels the contribution from the R family. In [13] two of us, discussing the importance of these cancellations in the context of the propagators, made the remark that a certain asymptotic relation between the second family and third family eigenvalues was a necessary condition for the cancellations to work. Although the asymptotic relation between the eigenvalues was certainly valid in the specific example discussed there, and may be valid in more general situations also, we can clearly see that it has nothing to do with the cancellations: these are a purely ”kinematic” effect, depending solely on the ultrametric geometry and on no further details of the theory. To conclude, we make an additional remark. Before presenting formulae (65)-(76) we gave a sketchy indication (with some details to be added in the Appendix) as to how they can be obtained, which is, of course, not necessarily the most economic way that they can be verified once known. Eqs. (65)-(76) are the inversion of (47)-(52) and of (41). The simplest way to check them is by direct substitution. Having established the inverse relations between the matrix elements and the kernel we can now summarise the steps one has to follow in order to invert an ultrametric matrix. First one has to determine the kernel and the replicon eigenvalues of the matrix by (41), (47)-(52). To get the replicon eigenvalues of the inverse matrix is trivial, they are the reciprocal of the original replicon eigenvalues. To obtain the new kernel requires the solution of (57). This is the hard core that remains to be cracked after the layer controlled by

ultrametricity has been peeled off. To find the new kernel requires the concrete knowledge of the matrix elements and as such it is outside the scope of the present paper. Assuming the new kernel has been found one finally obtains the elements of the inverse matrix via eqs. (65)-(76). Acknowledgement. We are grateful to M. M´ezard for persistently urging us to perform this work. Two of us (I.K. and T.T.) were partly supported by the Hungarian National Science Fund OTKA, grant No. 2090. References [1] M´ezard M, Parisi G and Virasoro M 1987 Spin Glass Theory and Beyond (Singapore: World Scientific) [2] M´ezard M and Parisi G 1990 J. Phys. A: Math. Gen.23 L1229 [3] M´ezard M and Parisi G 1991 J. Phys. I. France 1 809 [4] M´ezard M and Parisi G 1992 J. Phys. I. France 2 2231 [5] de Almeida J R L and Bruinsma R 1983 Phys. Rev. B35 7267 [6] M´ezard M and Young A P 1992 Europhys. Lett. 18 653 [7] Garel T and Orland H 1988 Europhys. Lett. 6 307 [8] Shaknovich E I and Gutin A M 1989 J. Phys. A: Math. Gen.22 1647 [9] Sasai M and Wolynes P G 1990 Phys. Rev. Lett. 65 2740 [10] Bouchaud J-P, M´ezard M and Yedidia J S 1991 Phys. Rev. Lett. 67 3840 [11] Parisi G 1980 J. Phys. A: Math. Gen.13 1887 [12] De Dominicis C, Kondor I and Temesv´ari T 1994 to be published in J. Phys. I. France [13] Kondor I and De Dominicis C 1986 Europhys. Lett. 2 617 [14] De Dominicis C and Kondor I 1984 J. Physique Lett. 45 L-205 [15] De Dominicis C and Kondor I 1985 J. Physique Lett. 46 L-1037 Appendix Our purpose here is to sketch the derivation of the replicon contributions to Eqs. (65-76) through what we called the representative vector. Eq. (58), written out in the original, Cartesian coordinates, reads: X f ˜ ij hj|γδi hαβ|iiM (A.1) Mαβ,γδ = ij

f are their biorthogonal where |ii now means any of the 21 n(n − 1) new basis vectors, and |ji counterparts. We are interested here in the contribution of the replicon family to (A.1) only, i.e. in R the partial sum, to be denoted by Mαβ,γδ , where i and j are restricted to the replicon Kr ˜ ˜ sector. But M is diagonal in that sector, Mij = λi δi,j , so we have X R Mαβ,γδ = hαβ|iiλi f hi|γδi. (A.2) i∈R

Now let α ∩ β = r. Then the replicon vectors |ii contributing to (A.2) have to be such that they have nonzero components on the r th level of the Parisi hierarchy. From the description of these vectors given in the main text we know, however, that their components on every other level are then identically zero, and furthermore even on the r th level their nonzero components are concentrated inside a single block of size pr ×pr . Evidently, the component (αβ) must belong to this block. Although we have not actually determined the biorthogonal f (and our purpose here is to show that we do not need to, either), it is obvious that set hi| f will share the above properties of |ii: it will have nonzero components on the r th level, hi| and there inside the same pr × pr block only. It follows that (γδ) must belong to the same block and α ∩ β = γ ∩ δ = r. With this we have identified the set of replicon vectors that, for a given α, β, γ, δ, R can give a nonzero contribution to Mαβ,γδ . This set can be decomposed into orthogonal classes, labelled by the triplet of integers r, k, l (k, l ≥ r + 1), as explained in Sec. 3. In a given class (r, k, l) we still have several nonorthogonal replicon vectors and, in principle, they all contribute to (A.2). What we wish to show is that, in fact, one can choose a single vector from each class (r, k, l) in such a way as to exhaust the contribution of the whole class. We have called this vector the representative vector of the class. The choice of the representative vector is not unique. It is best to choose it such that the component (γδ) belong to the ”darkest” block, where the vector has the components A. (Consult Figs. 5,6,7.) Now, as we have already pointed out in Sec. 3, the subspace orthogonal to the representative vector thus chosen is spanned by vectors that have zero components over this ”A-block”. In particular, if |r; k, li is the representative vector of the class (r, k, l) then the biorthogonal counterparts of all the other members of the class will lie in the space orthogonal to |r; k, li, hence their components in the ”A-block” where also the pair (γδ) resides must necessarily be zero and thus their scalar product with the unit vector |γδi will vanish. Therefore, the only contribution from the class (r, k, l) comes from the representative vector, indeed. The summation in (A.2) then runs over the representative vectors only, so we can rewrite (A.2) as R Mαβ,γδ

=

R+1 X

R+1 X

k=r+1 l=r+1

λ(r; k, l)hαβ|r; k, lihr;g k, l|γδi.

(A.3)

We now decompose hr;g k, l| into components parallel and orthogonal to |r; k, li. The orthogonal component will not contribute to (A.3) for the same reasons as above. So we are left with the parallel component only, which, in view of hr;g k, l|r; k, li = 1 and of the normalisation of |r; k, li, is nothing but the representative vector itself. So we can finally write R+1 X R+1 X R Mαβ,γδ = λ(r; k, l)hαβ|r; k, lihr; k, l|γδi. (A.4) k=r+1 l=r+1

Although the above consideration is quite trivial really, one may find it mystifying that it is possible to reconstruct a matrix from selecting a single vector from each class of

its eigenvectors which are nonorthogonal within the class. As it transpires from the proof, the key factor is that the subspace orthogonal to the representative vector is composed of vectors having vanishing components over the ”A-block”, and this in turn hinges upon the common property of all replicons, namely that the sum of their components in each row is zero. A little more reflection will show one, however, that the underlying reason is R that the matrix Mαβ,γδ does not depend on α, β, γ and δ separately, only on the various overlaps formed out of these indices, therefore the puzzling property of the representative vectors carrying all the information about M R can be directly linked to the ultrametric symmetries of M . R As an illustration of the use of (A.4), let us calculate the diagonal components Mαβ,αβ . There will be three kinds of terms contributing to (A.4): (i) k = l = r + 1: hr; r + 1.r + 1|αβi2 = A2 =

pr − 3pr+1 2 pr+1 (pr − pr+1 )

=

2µ(r; r + 1, r + 1) nδr

,

(A.5)

where use has been made of Eqs. (32), (33). (ii) k ≥ r + 2, l = r + 1: pr − 2pr+1 hr; k, r + 1|αβi = A = pr+1 (pr − pr+1 ) 2

2



1 1 − pk pk−1



=

2µ(r; k, r + 1) nδr

,

(A.6)

where we have used (34) and (35). (iii) k ≥ r + 2, l ≥ r + 2: 2

2

hr; k, l|αβi = A =



1 1 − pk pk−1



1 1 − pl pl−1



=

2µ(r; k, l) nδr

,

(A.7)

see (36), (37). Substituting (A.5), (A.6) and (A.7) back into (A.4) we find R Mαβ,αβ

=

r,r M R R+1,R+1

=

R+1 X

R+1 X

k=r+1 l=r+1

2µ(r; k, l) λ(r; k, l) nδr

which is precisely the second (replicon) term quoted in (65a). The replicon contributions to (66a), (67a), (74a), (75a) and (76a) can be worked out similarly.