No Arabic abstract
We present a methodology for ensuring the robustness of our analysis pipeline in separating the global 21-cm hydrogen cosmology signal from large systematics based on singular value decomposition (SVD) of training sets. We show how traditional goodness-of-fit metrics such as the $chi^2$ statistic that assess the fit to the full data may not be able to detect a suboptimal extraction of the 21-cm signal when it is fit alongside one or more additional components due to significant covariance between them. However, we find that comparing the number of SVD eigenmodes for each component chosen by the pipeline for a given fit to the distribution of eigenmodes chosen for synthetic data realizations created from training set curves can detect when one or more of the training sets is insufficient to optimally extract the signal. Furthermore, this test can distinguish which training set (e.g. foreground, 21-cm signal) needs to be modified in order to better describe the data and improve the quality of the 21-cm signal extraction. We also extend this goodness-of-fit testing to cases where a prior distribution derived from the training sets is applied and find that, in this case, the $chi^2$ statistic as well as the recently introduced $psi^2$ statistic are able to detect inadequacies in the training sets due to the increased restrictions imposed by the prior. Crucially, the tests described in this paper can be performed when analyzing any type of observations with our pipeline.
One of the last unexplored windows to the cosmos, the Dark Ages and Cosmic Dawn, can be opened using a simple low frequency radio telescope from the stable, quiet lunar farside to measure the Global 21-cm spectrum. This frontier remains an enormous gap in our knowledge of the Universe. Standard models of physics and cosmology are untested during this critical epoch. The messenger of information about this period is the 1420 MHz (21-cm) radiation from the hyperfine transition of neutral hydrogen, Doppler-shifted to low radio astronomy frequencies by the expansion of the Universe. The Global 21-cm spectrum uniquely probes the cosmological model during the Dark Ages plus the evolving astrophysics during Cosmic Dawn, yielding constraints on the first stars, on accreting black holes, and on exotic physics such as dark matter-baryon interactions. A single low frequency radio telescope can measure the Global spectrum between ~10-110 MHz because of the ubiquity of neutral hydrogen. Precise characterizations of the telescope and its surroundings are required to detect this weak, isotropic emission of hydrogen amidst the bright foreground Galactic radiation. We describe how two antennas will permit observations over the full frequency band: a pair of orthogonal wire antennas and a 0.3-m$^3$ patch antenna. A four-channel correlation spectropolarimeter forms the core of the detector electronics. Technology challenges include advanced calibration techniques to disentangle covariances between a bright foreground and a weak 21-cm signal, using techniques similar to those for the CMB, thermal management for temperature swings of >250C, and efficient power to allow operations through a two-week lunar night. This simple telescope sets the stage for a lunar farside interferometric array to measure the Dark Ages power spectrum.
We present an investigation of the horizon and its effect on global 21-cm observations and analysis. We find that the horizon cannot be ignored when modeling low frequency observations. Even if the sky and antenna beam are known exactly, forward models cannot fully describe the beam-weighted foreground component without accurate knowledge of the horizon. When fitting data to extract the 21-cm signal, a single time-averaged spectrum or independent multi-spectrum fits may be able to compensate for the bias imposed by the horizon. However, these types of fits lack constraining power on the 21-cm signal, leading to large uncertainties on the signal extraction, in some cases larger in magnitude than the 21-cm signal itself. A significant decrease in signal uncertainty can be achieved by performing multi-spectrum fits in which the spectra are modeled simultaneously with common parameters. The cost of this greatly increased constraining power, however, is that the time dependence of the horizons effect, which is more complex than its spectral dependence, must be precisely modeled to achieve a good fit. To aid in modeling the horizon, we present an algorithm and Python package for calculating the horizon profile from a given observation site using elevation data. We also address several practical concerns such as pixelization error, uncertainty in the horizon profile, and foreground obstructions such as surrounding buildings and vegetation. We demonstrate that our training set-based analysis pipeline can account for all of these factors to model the horizon well enough to precisely extract the 21-cm signal from simulated observations.
Maximally Smooth Functions (MSFs) are a form of constrained functions in which there are no inflection points or zero crossings in high order derivatives. Consequently, they have applications to signal recovery in experiments where signals of interest are expected to be non-smooth features masked by larger smooth signals or foregrounds. They can also act as a powerful tool for diagnosing the presence of systematics. The constrained nature of MSFs makes fitting these functions a non-trivial task. We introduce maxsmooth, an open source package that uses quadratic programming to rapidly fit MSFs. We demonstrate the efficiency and reliability of maxsmooth by comparison to commonly used fitting routines and show that we can reduce the fitting time by approximately two orders of magnitude. We introduce and implement with maxsmooth Partially Smooth Functions, which are useful for describing elements of non-smooth structure in foregrounds. This work has been motivated by the problem of foreground modelling in 21-cm cosmology. We discuss applications of maxsmooth to 21-cm cosmology and highlight this with examples using data from the Experiment to Detect the Global Epoch of Reionization Signature (EDGES) and the Large-aperture Experiment to Detect the Dark Ages (LEDA) experiments. We demonstrate the presence of a sinusoidal systematic in the EDGES data with a log-evidence difference of $86.19pm0.12$ when compared to a pure foreground fit. MSFs are applied to data from LEDA for the first time in this paper and we identify the presence of sinusoidal systematics. maxsmooth is pip installable and available for download at: https://github.com/htjb/maxsmooth
Measurement of the spatial distribution of neutral hydrogen via the redshifted 21 cm line promises to revolutionize our knowledge of the epoch of reionization and the first galaxies, and may provide a powerful new tool for observational cosmology from redshifts 1<z<4 . In this review we discuss recent advances in our theoretical understanding of the epoch of reionization (EoR), the application of 21 cm tomography to cosmology and measurements of the dark energy equation of state after reionization, and the instrumentation and observational techniques shared by 21 cm EoR and post reionization cosmology machines. We place particular emphasis on the expected signal and observational capabilities of first generation 21 cm fluctuation instruments.
21 cm power spectrum observations have the potential to revolutionize our understanding of the Epoch of Reionization and Dark Energy, but require extraordinarily precise data analysis methods to separate the cosmological signal from the astrophysical and instrumental contaminants. This analysis challenge has led to a diversity of proposed analyses, including delay spectra, imaging power spectra, m-mode analysis, and numerous others. This diversity of approach is a strength, but has also led to confusion within the community about whether insights gleaned by one group are applicable to teams working in different analysis frameworks. In this paper we show that all existing analysis proposals can be classified into two distinct families based on whether they estimate the power spectrum of the measured or reconstructed sky. This subtle difference in the statistical question posed largely determines the susceptibility of the analyses to foreground emission and calibration errors, and ultimately the science different analyses can pursue. In this paper we detail the origin of the two analysis families, categorize the analyses being actively developed, and explore their relative sensitivities to foreground contamination and calibration errors.