학술논문

Signal-preserving CMB component separation with machine learning
Document Type
Working Paper
Source
Subject
Astrophysics - Cosmology and Nongalactic Astrophysics
Astrophysics - Instrumentation and Methods for Astrophysics
Language
Abstract
Analysis of microwave sky signals, such as the cosmic microwave background, often requires component separation with multi-frequency methods, where different signals are isolated by their frequency behaviors. Many so-called "blind" methods, such as the internal linear combination (ILC), make minimal assumptions about the spatial distribution of the signal or contaminants, and only assume knowledge of the frequency dependence of the signal. The ILC is a minimum-variance linear combination of the measured frequency maps. In the case of Gaussian, statistically isotropic fields, this is the optimal linear combination, as the variance is the only statistic of interest. However, in many cases the signal we wish to isolate, or the foregrounds we wish to remove, are non-Gaussian and/or statistically anisotropic (in particular for Galactic foregrounds). In such cases, it is possible that machine learning (ML) techniques can be used to exploit the non-Gaussian features of the foregrounds and thereby improve component separation. However, many ML techniques require the use of complex, difficult-to-interpret operations on the data. We propose a hybrid method whereby we train an ML model using only combinations of the data that $\textit{do not contain the signal}$, and combine the resulting ML-predicted foreground estimate with the ILC solution to reduce the error from the ILC. We demonstrate our methods on simulations of extragalactic temperature and Galactic polarization foregrounds, and show that our ML model can exploit non-Gaussian features, such as point sources and spatially-varying spectral indices, to produce lower-variance maps than ILC - eg, reducing the variance of the B-mode residual by factors of up to 5 - while preserving the signal of interest in an unbiased manner. Moreover, we often find improved performance when applying our model to foreground models on which it was not trained.
Comment: 22 pages, 13 figures. v2 has some updated references