No Arabic abstract
21 cm Epoch of Reionization observations promise to transform our understanding of galaxy formation, but these observations are impossible without unprecedented levels of instrument calibration. We present end-to-end simulations of a full EoR power spectrum analysis including all of the major components of a real data processing pipeline: models of astrophysical foregrounds and EoR signal, frequency-dependent instrument effects, sky-based antenna calibration, and the full PS analysis. This study reveals that traditional sky-based per-frequency antenna calibration can only be implemented in EoR measurement analyses if the calibration model is unrealistically accurate. For reasonable levels of catalog completeness, the calibration introduces contamination in otherwise foreground-free power spectrum modes, precluding a PS measurement. We explore the origin of this contamination and potential mitigation techniques. We show that there is a strong joint constraint on the precision of the calibration catalog and the inherent spectral smoothness of antennae, and that this has significant implications for the instrumental design of the SKA and other future EoR observatories.
We discuss absolute calibration strategies for Phase I of the Hydrogen Epoch of Reionization Array (HERA), which aims to measure the cosmological 21 cm signal from the Epoch of Reionization (EoR). HERA is a drift-scan array with a 10 degree wide field of view, meaning bright, well-characterized point source transits are scarce. This, combined with HERAs redundant sampling of the uv plane and the modest angular resolution of the Phase I instrument, make traditional sky-based and self-calibration techniques difficult to implement with high dynamic range. Nonetheless, in this work we demonstrate calibration for HERA using point source catalogues and electromagnetic simulations of its primary beam. We show that unmodeled diffuse flux and instrumental contaminants can corrupt the gain solutions, and present a gain smoothing approach for mitigating their impact on the 21 cm power spectrum. We also demonstrate a hybrid sky and redundant calibration scheme and compare it to pure sky-based calibration, showing only a marginal improvement to the gain solutions at intermediate delay scales. Our work suggests that the HERA Phase I system can be well-calibrated for a foreground-avoidance power spectrum estimator by applying direction-independent gains with a small set of degrees of freedom across the frequency and time axes.
We analyse the accuracy of radio interferometric gridding of visibilities with the aim to quantify the Epoch of Reionization (EoR) 21-cm power spectrum bias caused by gridding, ultimately to determine the suitability of different imaging algorithms and gridding settings for 21-cm power spectrum analysis. We simulate realistic LOFAR data, and construct power spectra with convolutional gridding and w-stacking, w-projection, image domain gridding and without w-correction. These are compared against directly Fourier transformed data. The influence of oversampling, kernel size, w-quantization, kernel windowing function and image padding are quantified. The gridding excess power is measured with a foreground subtraction strategy, for which foregrounds have been subtracted using Gaussian progress regression, as well as with a foreground avoidance strategy. Constructing a power spectrum that has a bias significantly lower compared to the expected EoR signals is possible with the tested methods, but requires a kernel oversampling factor > 4000 and, when using w-correction, > 500 w-quantization levels. These values are higher than typical values used for imaging, but are computationally feasible. The kernel size and padding factor parameters are less crucial. Among the tested methods, image domain gridding shows the highest accuracy with the lowest imaging time. LOFAR 21-cm power spectrum results are not affected by gridding. Image domain gridding is overall the most suitable algorithm for 21-cm EoR experiments, including for future SKA EoR analyses. Nevertheless, convolutional gridding with tuned parameters results in sufficient accuracy. This holds also for w-stacking for wide-field imaging. The w-projection algorithm is less suitable because of the kernel oversampling requirements, and a faceting approach is unsuitable due to the resulting spatial discontinuities.
Precise instrument calibration is critical to the success of 21 cm Cosmology experiments. Unmitigated errors in calibration contaminate the Epoch of Reionization (EoR) signal, precluding a detection. Barry et al. 2016 characterizes one class of inherent errors that emerge from calibrating to an incomplete sky model, however it has been unclear if errors in the sky model affect the calibration of redundant arrays. In this paper, we show that redundant calibration is vulnerable to errors from sky model incompleteness even in the limit of perfect antenna positioning and identical beams. These errors are at a level that can overwhelm the EoR signal and prevent a detection. Finally, we suggest error mitigation strategies with implications for the Hydrogen Epoch of Reionization Array (HERA) and the Square Kilometre Array (SKA).
Our understanding of the intergalactic medium at redshifts $z=5$-$6$ has improved considerably in the last few years due to the discovery of quasars with $z>6$ that enable Lyman-$alpha$ forest studies at these redshifts. A realisation from this has been that hydrogen reionization could end much later than previously thought, so that large islands of cold, neutral hydrogen could exist in the IGM at redshifts $z=5$-$6$. By using radiative transfer simulations of the IGM, we consider the implications of the presence of these neutral hydrogen islands for the 21-cm power spectrum signal and its potential detection by experiments such as HERA, SKA, LOFAR, and MWA. In contrast with previous models of the 21-cm signal, we find that thanks to the late end of reionization the 21-cm power in our simulation continues to be as high as $Delta^2_{21}=10~mathrm{mK}^2$ at $ksim 0.1~h/$cMpc at $z=5$-$6$. This value of the power spectrum is several orders of magnitude higher than that in the conventional models considered in the literature for these redshifts. Such high values of the 21-cm power spectrum should be detectable by HERA and SKA1-LOW in $sim 1000$ hours, assuming optimistic foreground subtraction. This redshift range is also attractive due to relatively low sky temperature and potentially greater abundance of multiwavelength data.
We describe the validation of the HERA Phase I software pipeline by a series of modular tests, building up to an end-to-end simulation. The philosophy of this approach is to validate the software and algorithms used in the Phase I upper limit analysis on wholly synthetic data satisfying the assumptions of that analysis, not addressing whether the actual data meet these assumptions. We discuss the organization of this validation approach, the specific modular tests performed, and the construction of the end-to-end simulations. We explicitly discuss the limitations in scope of the current simulation effort. With mock visibility data generated from a known analytic power spectrum and a wide range of realistic instrumental effects and foregrounds, we demonstrate that the current pipeline produces power spectrum estimates that are consistent with known analytic inputs to within thermal noise levels (at the 2 sigma level) for k > 0.2 h/Mpc for both bands and fields considered. Our input spectrum is intentionally amplified to enable a strong `detection at k ~0.2 h/Mpc -- at the level of ~25 sigma -- with foregrounds dominating on larger scales, and thermal noise dominating at smaller scales. Our pipeline is able to detect this amplified input signal after suppressing foregrounds with a dynamic range (foreground to noise ratio) of > 10^7. Our validation test suite uncovered several sources of scale-independent signal loss throughout the pipeline, whose amplitude is well-characterized and accounted for in the final estimates. We conclude with a discussion of the steps required for the next round of data analysis.