Abstract
In the context of cosmic microwave background data analysis, we study the solution to the equation that transforms scanning data into a map. As originally suggested in "messenger" methods for solving linear systems, we split the noise covariance into uniform and nonuniform parts and adjust their relative weights during the iterative solution. With simulations, we study mock instrumental data with different noise properties, and find that this "cooling" or perturbative approach is particularly effective when there is significant low-frequency noise in the timestream. In such cases, a conjugate gradient algorithm applied to this modified system converges faster and to a higher fidelity solution than the standard conjugate gradient approach. We give an analytic estimate for the parameter that controls how gradually the linear system should change during the course of the solution.
Export citation and abstract BibTeX RIS
1. Introduction
In observations of the cosmic microwave background (CMB), map-making is an intermediate step between the collection of raw scanning data and the scientific analyses, such as the estimation of power spectra and cosmological parameters. Next generation CMB observations will generate much more data than those today, and so it is worth exploring efficient ways to process the data even though, on paper, the map-making problem has long been solved.
The time-ordered scanning data is summarized by
where d , m , and n are the vectors of time-ordered data (TOD), the CMB sky-map signal, and measurement noise. P is a sparse matrix in the time-by-pixel domain that encodes the telescope's pointing. Of several map-making methods (Tegmark 1997), one of the most common is the method introduced for the Cosmic Background Explorer (COBE; Janssen & Gulkis 1992). This optimal, linear solution is
where provides the standard generalized least squares minimization of the χ2 statistic,
Here we assume that the noise has zero mean 〈 n 〉 = 0, and the noise covariance matrix N = 〈 n n †〉 is diagonal in frequency space. In the case where the noise is Gaussian, the COBE solution is also the maximum likelihood solution.
With current computational power, we cannot solve for by calculating directly. The noise covariance matrix N is often sparse in the frequency domain and the pointing matrix P is sparse in the time-by-pixel domain. In experiments currently under design, there may be ∼1016 time samples and ∼109 pixels, so these matrix inversions are intractable unless the covariance is uniform (proportional to the identity matrix I). We can use iterative methods, such as conjugate gradient descent, to avoid the matrix inversions, and execute each matrix multiplication in a basis where the matrix is sparse, using a fast Fourier transform to go between the frequency and time domain.
As an alternative to conjugate gradient descent, Huffenberger & Næss (2018) showed that the "messenger" iterative method could be adapted to solve the linear map-making system, based on the approach from Elsner & Wandelt (2013) to solve the linear Wiener filter. This technique splits the noise covariance into a uniform part and the remainder, and introduces an additional vector that represents the signal plus uniform noise. This messenger field acts as an intermediary between the signal and the data and has a covariance that is conveniently sparse in every basis. Elsner & Wandelt (2013) also introduced a cooling scheme that takes advantage of the split covariance: over the course of the iterative solution, we adjust the relative weight of those two parts. Starting with the uniform covariance, the modified linear system gradually transforms to the final system, under the control of a cooling parameter. In numerical experiments, Huffenberger & Næss (2018) found that a map produced by the cooled messenger method converged significantly faster than for standard conjugate gradient methods, and to higher fidelity, especially on large scales.
Papež et al. (2018) showed that the messenger field approach is equivalent to a fixed-point iteration scheme, and studied its convergence properties in detail. Furthermore, they showed that the split covariance and the modified system that incorporates the cooling can be solved by other means, including a conjugate gradient technique, which should generally show better convergence properties than the fixed-point scheme. With simulations, we have confirmed this conclusion. In their numerical tests, Papež et al. (2018) did not find benefits to the cooling modification of the map-making system, in contrast to the findings of Huffenberger & Næss (2018).
In this paper, we show that the difference arose because the numerical tests in Papež et al. (2018) used much less low-frequency (or 1/f) noise than Huffenberger & Næss (2018), and show that the cooling technique improves map-making performance especially when the low-frequency noise is large. This performance boost depends on a proper choice for the pace of cooling. Kodi Ramanah et al. (2017) showed that for the Wiener filter the cooling parameter should be chosen as a geometric series. In this work, we give an alternative interpretation of the parameterizing process and show that for map-making the optimal choice (unsurprisingly) is also a geometric series.
In Section 2 we describe our methods for treating the map-making equation and our numerical experiments. In Section 3 we present our results. In Section 4, we list our conclusions. In the Appendix we derive the prescription for our cooling schedule.
2. Methods
2.1. Parameterized Conjugate Gradient Method
The messenger field approach introduced an extra cooling parameter λ to the map-making equation, and solved the linear system with an alternative parameterized covariance . The parameter represents the uniform level of (white) noise in the original covariance. The remainder is the nonuniform part of the original noise covariance. (Here N without any arguments denotes the original noise covariance matrix N = 〈 n n †〉.) In this work we find it more convenient to work with the reciprocal of cooling parameter η = λ−1, which represents the degree of heteroscedasticity (nonuniformity) in the parameterized covariance
When η = 1 this parameterized covariance N(η) equals N.
Papež et al. (2018) showed that the conjugate gradient method can be easily applied to the cooled map-making problem. In our notation, this is equivalent to iterating on the parameterized map-making equation
as we adjust the parameter through a set of levels {ηi }. (We use with no η argument to mean the estimated in Equation (2), independent of η.) This equation leads to the same system as the parameterized equation in the messenger field method, because N(η) = λ−1 N(λ) and the condition number do not change upon scalar multiplication to both sides of the equation. For concreteness we fix the preconditioner to M = P† P for all calculations.
When η = 0, the noise covariance matrix N(0) is homoscedastic (uniform), and the solution is given by the simple binned map , which can be solved directly.
Since the non-white part is the troublesome portion of the covariance, we can think of the η parameter as increasing the heteroscedasticity of the system, adding a perturbation to the solution achieved at a particular stage, building ultimately upon the initial uniform covariance model. Therefore, this quasi-static process requires η increase as 0 = η0 ≤ η1 ≤ ⋯ ≤ ηfinal = 1, at which point we arrive at the desired map-making equation, and the solution .
We may iterate more than once at each intermediate ηi : we solve Equation (5) with conjugate gradient iterations using the result from the previous calculation as the initial value. We move to the next parameter ηi+1 when the norm of the residual vector
is an order of magnitude smaller than the norm of the right-hand side of Equation (5).
This is not stringent enough to completely converge at this η-level, but we find that it causes the system to converge sufficiently to allow us to move on to the next η.
2.2. Analytical Expression for {ηi } Series
The next question is how to appropriately choose these monotonically increasing parameters η. We also want to determine η1, ⋯ ,ηn−1 before starting conjugate gradient iterations, because the time-ordered data d is very large, and we do not want to keep it in the system memory during calculation or repeatedly read it in from the disk. If we determine η1, ⋯ ,ηn−1 before the iterations, then we can precompute the right-hand side of Equation (5) for each ηi and keep these map-sized objects, instead of the entire time-ordered data.
In the Appendix, we show that a generic good choice for the η parameters is given by this geometric series
where are the eigenvalues of under frequency representation. This is one of our main results. It not only tells us how to choose parameters ηi , but also when we should stop the perturbation, and set η = 1. For example, if the noise covariance matrix N is almost uniform, then , and we would have . This tell us that we do not need to use the parameterized method at all, because η0 = 0 and η1 = η2 = ⋯ = 1. This corresponds to the standard conjugate gradient method with a simple binned map as the initial guess (as recommended by Papež et al. 2018).
2.3. Intuitive Interpretation of η
This section includes a way to interpret the role of η that is less technical than in the Appendix. Our ultimate goal is to find , which minimizes χ2( m ) in Equation (3). Since N is diagonal in frequency space, χ2 could be written as a sum of all frequency modes ∣( d − P m )f ∣2 with weight , such as . The weight is large for low-noise frequency modes (small Nf ), and small for high-noise modes, which means χ2( m ) would favor the low-noise modes, and therefore the conjugate gradient map-making focuses on minimizing the error *ε ≡ d − P m in the low-noise part.
After introducing η, we minimize χ2( m , η) in Equation (A1) instead. For η = 0, N−1(0) ∝ I the system is homoscedastic and the estimated map does not prioritize any frequency modes. As we slowly increase η, we decrease the weight for the high-noise modes, and focus on minimizing error for the low-noise part. If we start with η1 = 1 directly, which corresponds to the vanilla conjugate gradient method, then the algorithm will focus mostly on minimizing the low-noise part, such that χ2 would converge very quickly on the low-noise modes (typically high temporal frequencies and small spatial scales), but slowly on the high-noise modes (low frequencies and large scales). However, by introducing the η parameter, we let the solver first treat every frequency equally. Then as η slowly increases, it gradually gives more focus on the lowest noise part.
2.4. Computational Cost
To properly compare the performance cost of this method with respect to the vanilla conjugate gradient method with the simple preconditioner, we need to compare their computational cost at each iteration. We could define and , and Equation (5) could be written as . The right-hand side b (ηi ) could be computed before iterating, since we have determined ηi in advance, so it will not introduce extra computational cost. The most demanding part of the conjugate gradient method is calculating its left-hand side A(ηi ) m , because it contains a Fourier transform of P m from the time domain to frequency domain and an inverse Fourier transform of from the frequency domain back to time domain, which is order with n being the length of time-ordered data. Compared to the traditional conjugate gradient method, we swap N−1 with , and the cost is the same for one step, since both methods need a fast Fourier transform and inverse fast Fourier transform at one iteration.
At each ηi level, we use the residual to determine whether to switch to the next level (ηi+1), as in Equation (7). Calculation of the residual vector r ( m , ηi ) is part of the conjugate gradient algorithm, so this will not add extra cost either. Therefore, overall introducing the η will not have extra computational cost within the conjugate gradient iterations.
However, we start a new conjugate gradient algorithm whenever ηi updates to ηi+1. Thus we must reinitialize the conjugate gradient algorithm, recalculating the residual r ( m , ηi+1) based on new ηi+1. This residual calculation contains an extra A(ηi ) m operation. Therefore, if we have a series , there will have nη − 1 extra A(η) m operations compare to the traditional conjugate gradient method. If the total number of iterations is much larger than nη , then this extra cost is negligible. For our simulation, this extra step would have a rather significant impact on the final result. To have a fair comparison between the parameterized and traditional conjugate gradient method, we will present our results with the number of operations as the horizontal axis.
2.5. Numerical Simulations
To compare these algorithms, we need to do some simple simulations of scanning processes, and generate the time-ordered data from a random sky signal. 1 Our sky is a small rectangular area, with two orthogonal directions x and y, both with ranges from –1°to +1°. The signal has Stokes parameters (I, Q, U) for intensity and linear polarization.
For the scanning process, our mock telescope contains nine detectors, each with different sensitivity to polarization Q and U. It scans the sky with a raster scanning pattern. Its scanning frequency is fscan = 0.1 Hz and sampling frequency is fsample = 100 Hz. The telescope scans the sky horizontally then vertically. This gives the noiseless signal s . The sky signal in the timestream has an rms of 56 μK. The signal is continuous, so that it has structure on subpixel scales, but we find that our main conclusions remain the same when the input signal is pixelized. In map-making, we digitize the position (x, y) into 512 × 512 pixels.
We model the noise power spectrum with
which is white at high frequencies, a power law below the knee frequency, and gives us the option to flatten the low-frequency noise below an apodization frequency (like in Papež et al. 2018). Note that as fapo → 0, , and it becomes a 1/f-type noise model.
Dünner et al. (2013) measured the slopes of the atmospheric noise in the Atacama under different water vapor conditions, finding α = 2.7 to 2.9. Here we use σ2 = 10 μK2, α = 3, and compare the performance under different noise models. In our calculations, we choose different combinations of fknee and fapo as in Figure 1. The noise spectra with the most low-frequency noise have high fknee or low cutoff fapo.
The noise covariance matrix
is a diagonal matrix in frequency space, where Δf is equal to the reciprocal of total scanning time T ≈ 1.05 × 104 s.
Finally, we get the simulated time-ordered data d = s + n by adding up the signal and noise.
3. Results
We compare the standard conjugate gradient method versus the conjugate gradient with our perturbed linear system. Both methods use the simple preconditioner P† P. Figure 2 shows the χ2 results for 1/f noise models (fapo = 0) with different knee frequencies. Note that the χ2 values in all figures are calculated based on the standard χ2( m ) in Equation (3), not the η-dependent χ2( m , η) of the modified system (Equation (A1)). The minimum that we use for comparison is calculated from a deliberately slowed and well-converged parameterized conjugate gradient method: one with 100 η values and that halts when the final norm of the residual is smaller than , or 100 iterations after η = 1. From Figure 2, we can see for the 1/f noise model, when fknee ≳ 10fscan the parameterized method starts showing an advantage over the vanilla conjugate gradient method.
Download figure:
Standard image High-resolution imageIn Figure 3, we fixed fknee = 10 Hz, and change fapo. As we decrease fapo relative to fknee, increasing the amount of low-frequency noise, the parameterized conjugate gradient method performs better.
Download figure:
Standard image High-resolution imageLooking at the power spectrum in Figure 1, when fknee is small, or fapo is large, there is not much low-frequency noise. These situations correspond to the left-side plots in Figure 2 and Figure 3. The right-side graphs have a significant amount of low-frequency noise. We conclude that the introduction of the slowly varying η parameter improves performance most when there are large low-frequency noise contributions.
We also tried different 1/f noise slopes α. For α = 2, the conclusion is the same as α = 3. When α = 1, the low-frequency noise is reduced compared to the cases with steeper slopes, and the vanilla conjugate gradient method is preferred, except some cases with very large knee frequency like fknee = 100 Hz and fapo = 0, which favors the parameterized method. In Papež et al. (2018), the slope α = 1 and the noise power spectrum is flattened at fapo ≈ 0.1fknee. Their knee frequency is the same as their scanning frequency, so is most like our case when fknee = fscan = 0.1 Hz. Their case had little low-frequency noise, and we confirm their specific result that the standard conjugate gradient method converges faster in that case. In general, however, we find that cases with significantly more low-frequency noise benefit from the cooling/parameterized approach.
4. Conclusions
We analyzed the parameterized conjugate gradient map-making method that is inspired by the messenger field idea of separating the white noise out of the noise covariance matrix. Then we gave an analytical expression for the series of η parameters that govern how quickly the modified covariance adjusts to the correct covariance, and showed that this method adds only the extra computational cost of reinitializing the conjugate gradient process based on a new η parameter.
We tested this method for different noise power spectra, both flattened and nonflattened at low frequency. The results showed that the parameterized method is faster than the traditional conjugate gradient method when there is a significant amount of low-frequency noise. It could be further improved if we could get a more accurate estimation for the change in χ2 as a function of the η parameter, either before iteration or without using time-ordered data during iteration.
Also note that we fixed the preconditioner as M = P† P during our calculation, this parameterizing process could be applied to any preconditioner and possibly improve performance when there is a significant amount of low-frequency noise.
This type of analysis for the cooling parameter may also be apply to other areas, like the Wiener filter. Papež et al. (2018) showed that the messenger field method of Elsner & Wandelt (2013) for solving Wiener filter problem could also be written as a parameterized conjugate gradient algorithm. It stands to reason that such a system may also benefit from splitting and parameterizing its noise covariance, depending on the noise properties. (In the Wiener filter, Kodi Ramanah et al. 2017 additionally suggest that splitting the signal covariance and combining the uniform parts of the signal and noise.)
The benefits to map-making from a cooled messenger method seem to come from the cooling and not actually from the messenger field that inspired it. However, the messenger field approach may still have a role in the production and analysis of CMB maps. In particular, the close connection between the messenger method and Gibbs sampling may allow us to cheaply generate noise realizations of a converged map by generating samples from the map posterior distribution, something that we will continue to explore in future work.
For this work, B.C. and K.M.H. are supported by NSF award 1815887.
Appendix: The Derivation of η Parameter Series
We know that the initial degree of heteroscedasticity η0 = 0, which means the system is homoscedastic (uniform noise) to start. What would be a good value for the next parameter η1? To simplify notation, we use Nη to denote the parameterized covariance matrix . For some specific η value, the estimated map minimizes
with η being fixed. We restrict to the case that the noise covariance matrix N is diagonal in the frequency domain, and represent the frequency-domain eigenvalues as Nf .
The perturbative scheme works like this. We start with with , which could be solved directly. Then we use the conjugate gradient method to find and its corresponding . So let us consider η1 = η0 + δ η = δ η such that η1 = δ η is a very small quantity, δ η ≪ 1. (Remember η0 = 0.) Since minimizes χ2( m , η) with η being fixed, we have , and using the chain rule
Then the fractional decrease of from η0 to η1 = δ η is
Here we put a minus sign in front of this expression such that it is nonnegative, and use Nη=0 = τ I at the second equality. We want to be large to encourage fast convergence. Therefore, is much smaller than , or . Then we would expect
Here we use the notation 1−, which means the upper bound is close to but strictly smaller than 1. Now we can use Equation (A3) and let it equal 1; thus .
Applying this same idea to ηm+1 = ηm + δ ηm with m ≥ 1, we would get
As mentioned in the main text, we need to determine the entire series ηi before conjugate gradient iterations. We do not have the with which to calculate them and need to find another approach.
Let us go back to Equation (A3). Since we cannot calculate before making the map, we treat it as an arbitrary vector, then the least upper bound of Equation (A3) is given by
where is the maximum eigenvalue of . We want to be as large as possible, but it will not exceed 1. We combine Equations (A4) and (A6), and choose δ η such that the least upper bound is equal to 1, to make sure the process is not going too fast. Thus we have
Here Nf and are the eigenvalues of N and in the frequency domain. If the condition number of noise covariance matrix , then η1 ≈ κ−1(N).
What about the other parameters ηm with m > 1? We use a similar analysis, letting ηm+1 = ηm + δ ηm with a small δ ηm ≪ 1. First, let us find the least upper bound
The upper bound in the second line is a little bit tricky. Both matrix and can be simultaneously diagonalized in frequency space. For each eigenvector e f , the corresponding eigenvalue of the matrix on the numerator is , and the eigenvalue for the matrix in the denominator is . Their eigenvalues are related by . For any vector v = ∑f αf e f , we have
Again assuming , which we expect to be satisfied for ηm ≪ 1. That is because if η ≲ 1, would be close to the minimum χ2, which means , which would violate our assumption. Luckily, the final result (Equation (A14)) is a geometric series, only the last few ηm values fail to satisfy this condition. Similarly, we could set the least upper bound equal to 1. Then we get
Therefore
If written in the form it is easy to see that for m ≥ 1, forms a geometric series
where we used . Note that m = 0 and η0 = 0 also satisfy this expression and we have the final expression for all ηm
Here we need to truncate the series when ηm > 1.
In numerical simulations, we find that if we update the η parameter based on the more precise Equation (A5), there is only marginal improvement over the η series given in Equation (A14) for the 1/f noise model, and slight improvement when there is not much low-frequency noise. If we use Equation (A5), it ends up with fewer η parameters in the series, but the interval between ηi and ηi+1 gets larger. In our simulation this sometimes causes one more iteration at a certain η value, so in the end there are only slight improvements. For a large data set that needs a lot of iterations to converge from ηi to ηi+1, where several extra iterations may not be significant, using Equation (A5) may provide a larger performance boost.
Footnotes
- 1
The source code and other information are available at https://github.com/Bai-Chiang/CMB_map_making_with_cooling.