Abstract
For products PN of N random matrices of size , there is a natural notion of finite N Lyapunov exponents . In the case of standard Gaussian random matrices with real, complex or real quaternion elements, and extended to the general variance case for , methods known for the computation of are used to compute the large N form of the variances of the exponents. Analogous calculations are performed in the case that the matrices making up PN are products of sub-blocks of random unitary matrices with Haar measure. Furthermore, we make some remarks relating to the coincidence of the Lyapunov exponents and the stability exponents relating to the eigenvalues of PN.
Export citation and abstract BibTeX RIS
1. Introduction
Let Ai, be random matrices independently chosen from the same ensemble, and form the product
Under the condition that the second moment of the diagonal entries of are finite, the multiplicative ergodic theorem of Oseledec [22, 24] tells us that the limiting positive definite matrix
is well defined. In particular Vd has eigenvalues
with referred to as the Lyapunov exponents.
The recent study [3], building on [4], gave the asymptotic form of the eigenvalue distribution of in the case that each Ai was chosen from the ensemble of standard complex Gaussian matrices. For large N, and with the eigenvalues of , say, parametrized by writing , this was shown to be given by
where Sym denotes symmetrization in the stated variables and, with denoting the digamma function,
In the limit, it follows from this and the definitions in the first paragraph that the Lyapunov exponents are given by , , which is a known result [8]. The value of for i = 1 agrees with another known result [5]. One of the aims of the present paper is to show how for products of Gaussian random matrices can be computed by analogous calculations leading to [8, 18, 20]. We do this in section 2.2, and we point out there that the mean and variance of the finite N Lyapunov exponents for PN consisting of a product of Gaussian and inverse Gaussian matrices follows as a corollary.
The formulation used to perform the computations in section 2.2, which itself is presented in section 2.1, allows for the computation of the variances associated with the distribution of for large but finite N to also be computed in a number of other cases. One is for random matrices , where Σ is a fixed positive definite matrix and Gi a standard complex Gaussian random matrix. This is carried out in section 2.3. In section 2.4 we generalize this choice of Ai to the case that Gi is a standard real, or standard real quaternion, Gaussian random matrix, but for the variance corresponding to the largest Lyapunov exponent only. The Lyapanov exponents and associated variances are computed for products of sub-blocks of random unitary matrices with real, complex or real quaternion entries and chosen with Haar measure in section 3. We point out that the formulas for the Lyapunov exponents and the associated variances are analogous to those for products involving Gaussian random matrices of two different sizes in equal proportion, except for a sign.
The paper [3] also undertook a study of the eigenvalues of the matrix PN, say, parameterized by writing . It was found that for large N the joint distribution of is given by (1.4), with each yi replaced by , and multiplied by . Thus the angular dependence is uniform, while the asymptotic distribution of is the same as that for . On this latter point, following [13] as reported in the reference book [6, p 21], we argue that provided PN can be diagonalized, the Lyaponov exponents can be computed from either the eigenvalues of or the eigenvalues of PN itself. This point is made in section 4.2.
2. Computation of variances for Gaussian ensembles
2.1. Known results
While for a general ensemble of random matrices the Lyapunov spectrum is difficult to compute, there is a class of probability densities on the space of matrices—referred to as isotropic—for which the computation becomes relatively simple [5]. The defining feature of an isotropic matrix distribution on is that for a particular vector , has distribution independent of . Examples are when , where Σ is a fixed positive definite matrix and is a standard Gaussian matrix with real , complex or real quaternion entries. In the latter case, these entries are themselves matrices with complex entries, and the eigenvalues are calculated from the corresponding complex matrix which has a doubly degenerate spectrum; see e.g. [7, section 1.3.2].
The utility of an isotropic matrix distribution on can be seen by revising the formalism devised in [22, 24] for the computation of the Lyapunov spectrum. The main formula specifies the partial sum of Lyapunov exponents according to
where , the sup operation is over all sets of linearly independent vectors and refers to the volume of the parallelogram generated by the given set of k vectors. To quantify the latter, with
so that BN is the matrix with its columns given by the k vectors , we have
Suppose now that is a set of k linearly independent unit vectors, and let denote the matrix with 1's in the diagonal positions of the k rows, and 0's elsewhere. The fact that the distribution of is independent of tells us that
where denotes the matrix formed by the first k columns of , and it tells us furthermore that the sup operation is redundant. Consequently
while the law of large numbers gives that this can be evaluated as the average over
Computation of this average in the case implies [8, 18, 21]
Suppose (1.1) is generalized so that each Ai is of size , . It has recently been noted in [17] that the calculations leading to (2.6) generalize and we have
Note that (2.7) is a symmetric function of . This is in fact a general feature of the joint distribution of the eigenvalues of [15], and so without loss of generality we are free to suppose . The existence of the limit implies that the limit in (2.7) is well defined. Hence only a finite number of the matrices Ai can be rectangular, with all the rest being square. In [17] it was furthermore argued, using exact calculations based on knowledge of the joint distribution of the spectrum of PN [1, 2, 9, 15, 16] , that the asymptotic form of the eigenvalues in this setting is given by the form (1.4) with given by (2.7) and
(note that for —the case of complex entries—this reproduces the formula for in (1.5)). Our aim in the next section is to show how (2.8) can be deduced from analogous calculations as those leading to (2.6).
2.2. Variances for the standard Gaussian ensembles
Our starting point is the fact that (2.1) is valid for finite N, provided that the limiting operation is removed on the rhs, and with such that the eigenvalues of are equal to (see e.g. the discussion in [3, section 6.1]). In the Gaussian case this simplifies to (2.4) without the limiting operation on the rhs. The argument leading to (2.5) then gives that to leading order in N
The lhs can be rewriten
But for large N, and a non-degenerate Lyapunov spectrum, we expect the covariance to decay exponentially fast. This is in keeping with the joint probability density being of the form (1.4). Ignoring such terms, for large N we can substitute
for the lhs of (2.9). It thus remains to compute the rhs. In the case this is a straightforward task.
and thus the rhs of (2.9) is equal to
Proof. We follow the strategy of the proof given in [8, section 2.2] of the evaluation of the rhs of (2.5) in the case and .
The average in (2.11) is over standard Gaussian matrices with real , complex or real quaternion entries. Such matrices have probability density function proportional to , where in the case the trace operation takes only one of the independent eigenvalues (recall the eigenvalues in this case are doubly degenerate). In particular, the average is a function of , suggesting that we change variables by introducing the Wishart matrix . The corresponding Jacobian is proportional to
(see e.g. [7, equation (3.22)]). Making use too of the simple identity
we see that the lhs of (2.11) is equal to
This average is over positive definite Hermitian matrices with real (), complex () or real quaternion () entries, and distribution having a density function proportional to
We see that (2.14) is a function only of the eigenvalues of , suggesting that we change variables to the eigenvalues and the eigenvectors. The Jacobian—which has the feature that the eigenvalue and eigenvector terms factorize—is given in e.g. [7, proposition 1.3.4], and we deduce that (2.14) is equal to
where , and
The multidimensional integral (2.17) is familiar in random matrix theory as a limiting case of the Selberg integral, and as such is evaluated as a product of gamma functions (see e.g. [7, proposition 4.7.3]). This shows that (2.16) is equal to
Evaluating the derivative in terms of the digamma function gives (2.11).
If we now subtract the second term in (2.9), recalling (2.6), we see that the term in the final line of (2.11) cancels, leaving (2.12).□
Equating (2.10) as the evaluation of the lhs of (2.9) for large N with the evaluation (2.12) of the rhs we obtain
which is equivalent to (2.8) in the case of square matrices.
A minor modification of the above reasoning allows for the derivation of (2.8) in the case of rectangular matrices. An important feature is that the non-negative integer parameters () can only take on a finite number of different values. Let these values be and suppose they are in proportion where . The rhs of the formula (2.9) with then generalizes to read
Changing variables for we see that the exponent in the Wishart ensemble obtained from the corresponding Jacobian is now . The result of lemma 2.1 shows that (2.11) holds with d replaced by , and all quantities averaged over ν in the sense of (2.7).
A similar averaging of the mean and standard deviation holds in the case that PN consists of both Gaussian and inverse Gaussian matrices (see [10] for a study of the singular values in this setting in the complex case with N finite). From the definitions, if the Lyapunov exponents for products of Gaussian random matrices are , then the Lyapunov exponents for products of inverse Gaussian random matrices are . Equivalently, denoting the Lyapunov exponents for products of inverse Gaussian matrices by , we thus have
while for the associated variances we have
Consider now a matrix product PN with proportion Gaussian matrices and inverse Gaussian matrices, and denote the Lyapunov exponents and corresponding variances by , . Consideration of the derivation of (2.18), and making use of (2.19) and (2.20), we see that
2.3. Complex standard Gaussian ensembles with general Σ
It has been shown in [8] that for with a standard complex Gaussian random matrix and Σ a fixed positive definite matrix with eigenvalues of equal to , the Lyapunov spectrum corresponding to the product (1.1) is given by
This was possible because of the availability of an explicit formula for the average
Thus, from the works [12, 25] we can read off that this average is equal to (see e.g. [8, equation (2.21)])
The Harish-Chandra/Itzykson–Zuber matrix integral (see e.g. [7, proposition 11.6.1]), giving a determinant formula for a particular integral over the unitary group is an essential component of the derivation of (2.24). The analogous matrix integral over the orthogonal or symplectic group does not permit such a structured evaluation, so impeding further progress for calculating the Lyapunov spectrum of the Gaussian ensembles with general Σ and real or real quaternion entries. Differentiating (2.24) with respect to μ and setting gives a formula for the right-hand side of (2.9) in the case . This formula is equivalent to (2.22).
For purposes of calculating the variances associated with the Lyapunov spectrum, following the strategy of section 2.2 we require the analogous evaluation of
Moreover, according to (2.5) the second term in (2.9) is equal to , and we know from (2.22) the value of . Although the computation of (2.25) is elementary, simply requiring a double differentiation of (2.24) with respect to μ, unfortunately for general k the resulting formula is not very revealing, consisting of a number of terms involving either summations or double summations up to k, with some of the summands being determinants. We will therefore restrict attention to the cases k = 1, when there is much simplification.
Proposition 2.2. Let be the eigenvalues of , assumed distinct. The average (2.25) in the case k = 1 is equal to
It follows that for large N
Proof. Our task is to differentiate (2.24) with k = 1 twice with respect to μ and to set . Since the only dependence on ν for k = 1 is in the first row, this reduces to computing the second derivative of each term in the first row. Noting
and substituting for the first row of (2.24) we see after some simple manipulation that the sought average is equal to
where in eliminating the determinant which would otherwise appear on the second line use has been made of the Vandermonde determinant evaluation
Recalling (2.22) in the case k = 1 allows (2.30) to be equivalently written
Expanding the determinants by the first row and making use of (2.29) to evaluate the cofactors we obtain (2.26).
To deduce (2.27) from (2.26) we subtract in keeping with (2.9) and equate with (2.10) in the case k = 1.□
2.4. The variance for Gaussian ensembles with general Σ
Our derivation of the formula (2.27) for in the case of general variance complex Gaussian matrices has made use of the determinant formula (2.24), which as already mentioned has no analogue in the real or real quaternion cases. This was similarly the case for the derivation given in [8] of the Lyapunov exponents (2.22). However, with regards to the latter with k = 1 (largest Lyapunov exponent) it has recently been shown by Kargin [18] that an alternative calculation is possible, which furthermore does generalize to the real and real quaternion cases, giving the formula [18, theorem 1.1]
In (2.31) γ denotes Euler's constant and for J true and otherwise. Here we adapt the method of [18] to the computation of the rhs of (2.9) with k = 1.
Proposition 2.3. Let denote the simple closed contour starting at large , goes along the upper edge of the real axis to , where , crosses to the lower edge of the real axis and returns to R. Define
We have
Proof. For k = 1 the quantity is a positive scalar random variable. Let its density function be denoted . We then have that
It is shown in [18] that
where is as in (2.32). Substituting (2.35) in (2.34), interchanging the order of integration and noting
shows that
where
By deforming the contour to go around the pole at z = 0 we deduce that . We read off from [18, equation (13)] that
Also, simple manipulation shows
where Jp is specified by (2.32). We substitute (2.38) and (2.39) in (2.36) then subtract , using the fact that [18, equation (14)]
to arrive at (2.33).□
In the case the integrals in (2.33) can be evaluated using the residue theorem to reclaim (2.27), noting too that . In the case that and thus , provided use of the residue theorem shows that
which when substituted in (2.33) can be verified to give the same value of as (2.8) in the case i = 1. It is shown in [18] that
as is consistent with the equality between (2.31) and (2.40). Analogous working shows
so it is also possible to express (2.33) in terms of real integrals.
3. Products of random truncated unitary matrices
3.1. Isotropic property
As remarked in the opening paragraph of section 2.1, the tractability of the computation of the Lyapunov spectrum for the Gaussian random matrices G is due to Gaussian matrices belonging to the class of isotropic matrices, meaning that any vector , has distribution independent of . This property holds for any class of random matrices that are unitary invariant, and thus their distribution in unchanged by multiplication by a unitary matrix U, which is furthermore required to have entries of the same type (real, complex or real quaternion) as the random matrices.
Let be a random unitary matrix of size with real (, complex () or real quaternion () elements, and chosen with Haar measure. The latter has the defining feature that it is unchanged by multiplication of the matrices by fixed unitary matrices of the appropriate type. Suppose the fixed unitary matrices have the block structure
where is a unitary matrix. Such matrices act only on the top sub-matrix of , so it must be that square submatrices of Haar distributed unitary matrices (also referred to as truncated unitary matrices [26]) are themselves unitary invariant, and are thus examples of isotropic matrices. With the matrices Ai in (1.1) chosen from the ensemble of truncated unitary matrices, our interest in this section is to compute the Lyapunov spectrum as well as the associated variances. Crucial to our computations is the fact that a () sub-block is distributed with a probability density function proportional to
We remark that the singular values of products of truncated complex unitary matrices for finite N are the topic of the recent work [19].
3.2. Computation of the Lyapunov spectrum and variances
According to (2.5) we have
where the average is with respect to (3.1). The rhs of (3.2) is easy to compute, allowing for a simple formula for the individual .
Proposition 3.1. Let the matrices Ai in (1.1) be chosen as the top sub-block of a Haar distributed random unitary matrix with real , complex or real quaternion entries. For at least we have
Proof. As in the Gaussian case, an essential point is that the average (3.2) is a function only of the positive definite matrix . Since is a sub-matrix of a unitary matrix, is further constrained to have eigenvalues less than 1, which we denote by writing . Regarding this as a change of variables, the corresponding Jacobian is proportional to (2.13). Hence our task becomes that of evaluating
where has distribution with a probability density function proportional to
with c2 as in (3.1) and
Both (3.4) and (3.5) are functions of the eigenvalues of only. Introducing the appropriate Jacobian, as can be found in e.g. [7, proposition 1.3.4], gives for (3.4) the expression in terms of multiple integrals
where
This is precisely the Selberg integral (see e.g. [7, chapter 4]), already seen in a limiting form in (2.17). Inserting its gamma functions evaluation reduces (3.7) to
Equating this with the lhs of (3.2) and inserting the values of c1 and c2 from (3.6) and (3.1) we arrive at (3.3).□
We remark that even though our derivation of (3.3) requires , used for example in the convergence of the integrals in (3.7), the final expression is well defined for , and we expect it to remain valid. A similar effect holds in other calculations relating to truncated unitary matrices which are based on (3.5) [11]. As evidence for our claim, note that for n = 0, (3.3) gives that each Lyapunov exponent is equal to 0. This is consistent with the matrices in the product (1.1) then themselves being unitary matrices, as the sub-matrix of is for n = 0 the matrix itself. Furthermore, for specific the predictions given by (3.3) are found to be consistent with numerical computations, using the methods of section 4.1 below (for example with n = 1, d = 2 a Monte-Carlo estimate with trials gave while (3.3) gives ).
Similar working gives the variances associated with the Lyapunov exponents (3.3).
Proposition 3.2. The analogue of the formula (2.8) for the variances associated with the Lyapunov exponents in the case of products of the truncated unitary matrices of proposition 3.1 is given by
Proof. We use the analogue of the formula (2.9) with the lhs replaced by its leading large N form (2.10). Thus our main task is to compute
Following the working in the proof of proposition 3.1 we see that this average is equal to
We recognize from (3.9) that the final line is equal to , and so cancels out of the rhs of (2.9). Thus
Recalling the values of c1 and c2 from (3.6) and (3.1) we see that this is equivalent to the stated formula (3.10). □
The formulas (3.3) and (3.10) have some similarity with the corresponding formulas (2.7) and (2.8) for products of rectangular Gaussian random matrices, one of size (and thus square), the other of size . In particular the formulas agree if the matrices were to occur in 'proportion' 1, and the matrices in 'proportion' −1.
4. Discussion
4.1. Comparison with numerical experiments
In our earlier paper on Lyapunov exponents [8] a discussion and references were given on the difficulty of accurately computing the Lyapunov exponents. A more robust, but less accurate approach is to use Monte-Carlo simulation based on (2.1). Numerically stable ways to carry out such computations have been reported in [6]. Thus, in relation to , and with d = 2, , we iteratively define , where . According to (2.1), a Monte-Carlo estimate of the largest Lyapunov exponent is then given by
and similarly an estimate of the corresponding variance is given by
To use (2.1) to compute , again taking d = 2 for simplicity, let and , then define , (, ) where is obtained from by the Gram–Schmidt orthonormalization procedure. It follows from (2.1) that a Monte-Carlo estimate of is
where Xi is the matrix with columns given by and , while an estimate of the corresponding variance is
As an example, let's consider the case of (1.1) with . We choose the eigenvalues of to equal so that
The formula (2.27) gives , whereas a Monte-Carlo calculation based on (4.2) with gave .
To perform Monte-Carlo simulations in relation to products of truncated unitary matrices, we must first generate random unitary matrices with Haar measure. One way to do this is to apply the Gram–Schmidt orthogonalization producure to a standard Gaussian matrix. In the complex case, with we did this and with found the estimates and whereas according to proposition 3.1 and . For the corresponding variances we found and whereas proposition 3.2 gives and .
4.2. Asymptotic spectrum of PN
As revised in the first paragraph of the Introduction, the Lyapunov exponents are defined in terms of the eigenvalues of the positive definite matrix , where PN is the random matrix product (1.1). In the recent works [3, 17], for the particular cases that the Ai in (1.1) are standard Gaussian matrices with real, complex or real quaternion entries, explicit asymptotic analysis of the joint distribution of the spectrum of PN has revealed that with the eigenvalues written in the form , the —referred to as the stability exponents—are identical to the Lyapunov exponents. Moreover in [3] a proof of the equality between the Lyapunov and stability exponents in the case of random matrices from an isotropic ensemble has been given. It was then conjectured that all matrix products formed from an isotropic ensemble have this property.
Actually the question of the asymptotic equality between the Lyapunov and stability exponents has been investigated at least as far back as 1987 [13], albeit in the setting of the linearization of sets of differential equations about fixed points. In this work it was conjectured that the exponents agree whenever the spectrum of PN is non-degenerate. The reference book [6] cites [13], and furthermore addresses directly the case of products of random matrices, writing 'if the matrix PN can be written in diagonal form, then (in our notation) '. One immediate consequence of this conjecture is the prediction of the Lyapunov spectrum for a product of upper triangular random matrices. There, up to ordering where Akk denotes the entry in the matrices Ai of (1.1), so the conjecture gives this same set for the Lyapunov exponents. In the case of this agrees with a known theorem [23].
Following [14], we emphasize that the conjecture appears to be an intrinsic property of the asymptotic limit of random matrix products, and is not shared by other asymptotic limits of spectra in random matrix theory. For example, in the case of standard real, complex or real quaternion standard Gaussian random matrices
where denotes the eigenvalue of largest modulus of Y. That this ratio is at least unity follows from the general inequality , which in turn is a consequence of the sub-multiplicity of the rhs viewed as a matrix norm.
Acknowledgments
This work was supported by the Australian Research Council for the project DP140102613. I thank Jesper Ipsen for comments on the first draft.