Abstract

The explicit representation for the limiting spectral moments of sample covariance matrices generated by the periodic autoregressive model (PAR) is established. We propose to use the moment-constrained maximum entropy method to estimate the spectral density function. The experiments show that the maximum entropy spectral density function curve obtained based on the fourth-order limiting spectral moment can match histograms of the eigenvalues of the covariance matrices very well.

1. Introduction

Periodic time-series models are used for modeling stationary time series with periodic characteristics. Periodicity refers to wave or oscillating movements around long-term trends presented in the time series, such as the water sector, where the supply of water varies with seasonal rainfall. In the economic time series, it is often accompanied by some cycle or quasi-periodic fluctuations, such as day and night changes of urban electricity consumption, seasonal changes of railway passenger transport, and repeated acceleration, interruption, stagnation, and recovery of economic growth; see Franses and Paap [1], Bell et al. [2], and Aliat and Hamdi [3]. The periodic autoregressive (PAR) model is a generalisation of the classical autoregressive (AR) model by allowing the parameters to vary with the variables over a period:

For a given and a predetermined value , the random variable denotes the value in the quarter of the cycle, where and . In the cycle, is an AR process. The model order is given by , and , represents the AR model coefficients for the quarter. The error process corresponds to a periodic white noise, with and . Let , where the transpose (indicated by ) of a row vector is a column vector, and we define the large-dimensional sample covariance matrices as , with the periodic length tends to infinity in proportion to the sample size , namely, . Sample covariance matrices are very important in multivariate statistical inference since many test statistics are defined by their eigenvalues or functions.

Random matrix theory (RMT) originated from the study of the energy levels of a large number of particles in quantum mechanics. Many of the laws of mathematical physics were discovered through numerical studies. In the late 1950s, E. P. Wigner formulated the problem in terms of the empirical distribution of random matrices [4, 5], which led to the study of the semicircular law of Gaussian matrices. Since then, RMT has formed an active branch in modern probability theory. The LSD of the sample covariance matrix of a large-dimensional random matrix composed of independent random variables, the MP law, was first established by Marcenko and Pastur [6]. Further research efforts were conducted to estimate the LSD of a product of two random matrices. Yin and Krishnaiah [7] proved the existence of LSD for the matrix sequence when the two matrix sequences and are standard Wishart matrices and positive definite matrices, respectively. In particular, Bai et al. [8] obtained a display representation of the LSD of the matrix sequence when the two matrix sequences and are the sample covariance matrix and the Wigner matrix, respectively. To relax the independence of the entries of , Silverstein [9] considered the case of , where is a nonnegative definite matrix and consists of independent and identically distribution (i.i.d.) entries, and the sample covariance matrices take the form of . Random matrices of the form were extensively investigated by many researchers, including Yin [10], Silverstein [9], Bai and Zhou [11], Jin et al. [12], Zhang et al. [13] and Yao [14]. Silverstein’s important work [9] aimed to relax the independence structure between coordinates and considered random vectors of the form , assuming that the spectral norm of the matrix sequence is bounded and that the ESD of the matrix sequence is convergent. Silverstein proved that the LSD of the sample covariance matrix exists, given by the LSD of equation representation of the Stieltjes transformation.

This paper investigates the LSD of large-dimensional sample covariance matrices generated by the PAR model, and the sample covariance matrices take the form of . A similar ARMA-type processes problem has been considered by Bai and Zhou [11], Jin et al. [12], Yao [14], etc. It follows that, in the ARMA-type processes, the covariance matrices equals to the Toeplitz matrix, according to the fundamental eigenvalue distribution theorem for Toeplitz-type matrices [15], the ESD of weakly converges to the nonrandom distribution H. For some situations, the ESD of have explicit form; then, the LSD for the sample covariance matrix can be solved explicitly. For the random matrix generated by the first-order vector autoregressive (VAR(1)) model and the first-order vector moving average (VMA(1)) model, Jin et al. [12] and Bai and Zhou [11] obtained a display representation of this LSD. However, in our first-order variable-coefficient PAR process, we consider the process be interrupted by certain factors that the odd terms and the even terms of the coefficient converge to and , respectively, and that . The mixing matrix is a lower triangular matrix whose main diagonal elements are all 1 and subdiagonal elements differ from each other. When , the limiting form of the matrices are Toeplitz matrices, and the eigenvalues of Toeplitz matrices have been proved by Gregory [16]; when , the ESD of has no explicit form, see [17, 18], and the structure of the limiting matrices are very complex. Our main contributions in this paper include the following. (i) For the first-order variable-coefficient PAR model, the explicit forms of limiting spectral moments for are given; through verifying the Carleman condition, we derive the explicit spectral moments of . (ii) A framework of the maximum entropy-based method is provided for estimate the LSD.

The paper is organized as follows. In Section 2, we employ the moment method to deduce the spectral density of sample covariance matrices. In Section 3, we use the maximum entropy method to find the approximate solution and analyze the convergence of the maximum entropy solution. We simulate the PAR model to obtain the periodic sample covariance matrices, and the experiment shows that the maximum entropy spectral density function curve can match histograms of the eigenvalues of the covariance matrices very well.

2. Methodology and Main Results

In this section, we first derive the explicit spectral moments form of ; subsequently, we verify the spectral moments statistics satisfying the Carleman condition, and then, we employ the moment method to deduce the spectral density of sample covariance.

We consider first-order PAR model:

Let

Then, the first-order PAR model can be defined aswhere denotes the equation on the left can push the equation on the right. From equation (3), is a lower triangular matrix whose main diagonal elements are all 1, and therefore, the matrix is invertible. Hence,

The sample covariance matrices and have the same ESD. Let , and the ESD of is the same as that of . Next, we establish the existence of the LSD of the covariance matrix sequence .

Theorem 1. Let be generated by a first-order PAR model in equation (2), and the odd terms and the even terms of the coefficient converge to and , respectively, and . The error is independent random variables satisfying for any , ; furthermore, we assume that ; then, the spectral moments of sample covariance matrix tend towhere is the limit of spectral moments of , which have an explicit form:

Proof. The proof is divided into two steps. We first show the existence of the limit of spectral moments which is defined as ; then, we verify the Carleman condition. From equation (3), the inverse of matrix can be written aswhere can be written asLetThen,Furthermore, we obtain thatApplying the trace operator, we haveBased on the assumption that , we obtainand therefore,LetThen,The following section wishes to obtain a simplified form of the function . From equations (10) and (16), we haveDefineThen, can be rewritten asLetTherefore, can have a simplified representation asBy the derivation operations,we obtain thatBy partially replacing in equation (24) with , we define a new functionDefineby whichsoThen,We use to take place of and obtain thatTherefore,Next, we verify the Carleman condition .
From equation (30), it follows thatThen, applying the equality and , we obtainFrom Stirling’s formula , we get . Then, ; applying this inequality, has a finite lower bound:Then,With the moment convergence theorem, the spectral distribution of tends to a nonrandom probability distribution H. Let and ; by the important Theorem 2.10 of Bai [19], the ESD of tends to a nonrandom limit distribution in probability (or almost surely). The spectral moments of satisfy the following equation of Yin [10]:From the moment convergence theorem, we can obtain that the ESD of converges with probability close to 1 to a nonrandom distribution, determined by the spectral moment statistic.

Remark 1. By the proof of Theorem 1, the limiting spectral moments of depend only on the limiting coefficient and . When , the limiting form of the matrices are Toeplitz matrices, and the eigenvalues of Toeplitz matrices have been proved, see [16]. When , the limiting matrices of are tridiagonal matrices and take the form like the following:The eigenvalues of this limiting matrix have no explicit form, see [18] and [17]. The limiting form of the matrices are very complex, and we obtain the explicit spectral moments of .

3. Estimating Spectral Density Function with Maximum Entropy

Unfortunately, there are no explicit LSD forms of this sample covariance matrices . In this section, with the spectral moments of , we employ the maximum entropy method to estimate the LSD. Mead and Papanicolaou [20] studied the maximum entropy method to solve the uncertainty problem. Granziol et al. [21] demonstrated the equivalence between the MaxEnt principle and the optimization of variables with constraints. Gao and Han [22] used the maximum entropy method to solve nonlinear expectation. Every possible event has a probability of occurring, information entropy is related to all possibilities, and a system with the highest information entropy is most likely, when all known information, taken into account. In mathematics, the variable is generally estimated by maximising the Shannon entropy of the information of the moments of the random variable of the density function:where is the support of the distribution. In this paper, the sample covariance matrices is nonnegative definite, so the estimated limiting distribution should be supported on .

Given the first N + 1 moments , the idea of determining the density function of a probability distribution based on the principle of maximum entropy is to find the density function such thatwhere .

Lagrange multiplier is a method for finding the extremes of a multivariate function under a set of constraints. With the help of the Lagrange multiplier method, the entropy function is defined as

When the first-order derivative of the function H with respect to and are equal to zero, the constraint defined in equation (39) is returned; then, the first-order derivative of the function do not provide any additional information. The second of these derivatives, however, evaluate to

The general solution of equation (41) iswhere we have set .

We obtain the maximum entropy solution by solving for the Lagrange multiplier . To do this, we must solve the following nonlinear equation:

When , we can obtain a display representation of the Lagrange multiplier. Taking N = 2 as an example, from equation (36), we have

Based on equation (40), we obtain as

Then,

When , we cannot obtain a display solution for the Lagrange multiplier , but at this point, the numerical solution for can be calculated using numerical optimization algorithms. From equation (40), the analytic expressions of the gradient is

If we denote the maximum entropy solution obtained by the method based on moments asthen the Hessian matrix is given by

is the kth moment of the simulated distribution, and the Newton iteration is done by searching until the termination criterion holds. We obtain the spectral density through Algorithm 1.

Input:, Ending accuracy , Diitter variance in Hessian ;
Output: Coefficients;
 initial , k = 0;
repeat
  Compute gradient : ;
  Compute Hessian ;
  Direction of Newton’s law: ;
  Next iteration point: ;
  k = k+1;
until.

We can obtain that the maximum entropy solution is weakly convergent, i.e.,

In our simulations, we use the Newton method from the Python-based SciPy package to solve for minimisation . It computes the gradient within a ending accuracy as well as the Hessian matrix. To give the Hessian matrix a faster rate of convergence, we symmetrize it and add a disturbance of intensity along the diagonal. In the simulation of the algorithm, we observe that the algorithm is not too sensitive to these choices.

From equations (31) and (36), we obtain the four limiting spectral moments of the sample covariance matrices as follows:

The authors of [23, 24] state that, as the number of moments increases (especially when ), the maximum entropy method becomes sensitive to numerical differences between the order moments, the gradient, and the state of the Hessian matrix, as well as improper arithmetic accuracy becomes very sensitive and can be improved by introducing different basis functions, such as a drift with Chebyshev polynomial basis functions with drift to overcome these effects. In this section, to simplify the operations, we estimate the maximum entropy density function in terms of 4th-order spectral moments. The value of (d; n; a; b) determines the value of each spectral moment. Figures 16 show histograms of eigenvalues and maximum entropy spectral density curves (MaxEntPDF) for different values of (d; n; a; b). From these plots, it can be seen that the maximum entropy spectral density function curve can match histograms of the eigenvalues of the covariance matrices very well.

4. Conclusion

In this paper, we investigate the LSD of large-dimensional sample covariance matrix generated by the first-order variable-coefficient PAR process. We consider the process be interrupted by certain factors that the odd terms and the even terms of the coefficient converge to and , respectively, and that . The main contributions in this paper include the following. (i) For the first-order variable-coefficient PAR model, the explicit form of limiting spectral moments for is given; through verifying the Carleman condition, we derive the explicit spectral moments of . (ii) A framework of the maximum entropy-based method is provided for estimate the LSD. We have some ideas for the next step of the research. On the one hand, when generalizing the convergence of the coefficients of the PAR model to a more general structure, how about obtaining an expression for the spectral moments of the limiting covariance matrix? On the other hand, it is of importance to extend the method presented in this paper to PAR with larger order.

Data Availability

No data were used to support this study.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The work was supported by National Natural Science Foundations of China under Grant 11531001.