Abstract
The fluctuations in the quantum spectrum could be treated like a time series. In this framework, we explore the statistical self-similarity in the quantum spectrum using the detrended fluctuation analysis (DFA) and random matrix theory (RMT). We calculate the Hausdorff measure for the spectra of atoms and Gaussian ensembles and study their self-affine properties. We show that DFA is equivalent to the statistics of RMT, unifying two different approaches. We exploit this connection to obtain theoretical estimates for the Hausdorff measure.
- Received 15 July 2005
DOI:https://doi.org/10.1103/PhysRevE.73.015201
©2006 American Physical Society