No Arabic abstract
Upcoming surveys such as LSST{} and Euclid{} will significantly improve the power of weak lensing as a cosmological probe. To maximise the information that can be extracted from these surveys, it is important to explore novel statistics that complement standard weak lensing statistics such as the shear-shear correlation function and peak counts. In this work, we use a recently proposed weak lensing observable -- weak lensing voids -- to make parameter constraint forecasts for an LSST-like survey. We use the cosmoslics{} $w$CDM simulation suite to measure void statistics as a function of cosmological parameters. The simulation data is used to train a Gaussian process regression emulator that we use to generate likelihood contours and provide parameter constraints from mock observations. We find that the void abundance is more constraining than the tangential shear profiles, though the combination of the two gives additional constraining power. We forecast that without tomographic decomposition, these void statistics can constrain the matter fluctuation amplitude, $S_8$ within 0.3% (68% confidence interval), while offering 1.5, 1.5 and 2.7% precision on the matter density parameter, $Omega_{rm m}$, the reduced Hubble constant, $h$, and the dark energy equation of state parameter, $w_0$, respectively. These results are tighter than the constraints from the shear-shear correlation function with the same observational specifications for $Omega_m$, $S_8$ and $w_0$. The constraints from the WL voids also have complementary parameter degeneracy directions to the shear 2PCF for all combinations of parameters that include $h$, making weak lensing void statistics a promising cosmological probe.
Cosmic voids are an important probe of large-scale structure that can constrain cosmological parameters and test cosmological models. We present a new paradigm for void studies: void detection in weak lensing convergence maps. This approach identifies objects that relate directly to our theoretical understanding of voids as underdensities in the total matter field and presents several advantages compared to the customary method of finding voids in the galaxy distribution. We exemplify this approach by identifying voids using the weak lensing peaks as tracers of the large-scale structure. We find self-similarity in the void abundance across a range of peak signal-to-noise selection thresholds. The voids obtained via this approach give a tangential shear signal up to $sim40$ times larger than voids identified in the galaxy distribution.
We propose counting peaks in weak lensing (WL) maps, as a function of their height, to probe models of dark energy and to constrain cosmological parameters. Because peaks can be identified in two-dimensional WL maps directly, they can provide constraints that are free from potential selection effects and biases involved in identifying and determining the masses of galaxy clusters. We have run cosmological N-body simulations to produce WL convergence maps in three models with different constant values of the dark energy equation of state parameter, w=-0.8, -1, and -1.2, with a fixed normalization of the primordial power spectrum (corresponding to present-day normalizations of sigma8=0.742, 0.798, and 0.839, respectively). By comparing the number of WL peaks in 8 convergence bins in the range of -0.1 < kappa < 0.2, in multiple realizations of a single simulated 3x3 degree field, we show that the first (last) pair of models can be distinguished at the 95% (85%) confidence level. A survey with depth and area (20,000 sq. degrees), comparable to those expected from LSST, should have a factor of approx. 50 better parameter sensitivity. We find that relatively low-amplitude peaks (kappa = 0.03), which typically do not correspond to a single collapsed halo along the line of sight, account for most of this sensitivity. We study a range of smoothing scales and source galaxy redshifts (z_s). With a fixed source galaxy density of 15/arcmin^2, the best results are provided by the smallest scale we can reliably simulate, 1 arcminute, and z_s=2 provides substantially better sensitivity than z_s< 1.5.
In this paper, we show that Minkowski Functionals (MFs) of weak gravitational lensing (WL) convergence maps contain significant non-Gaussian, cosmology-dependent information. To do this, we use a large suite of cosmological ray-tracing N-body simulations to create mock WL convergence maps, and study the cosmological information content of MFs derived from these maps. Our suite consists of 80 independent 512^3 N-body runs, covering seven different cosmologies, varying three cosmological parameters Omega_m, w, and sigma_8 one at a time, around a fiducial LambdaCDM model. In each cosmology, we use ray-tracing to create a thousand pseudo-independent 12 deg^2 convergence maps, and use these in a Monte Carlo procedure to estimate the joint confidence contours on the above three parameters. We include redshift tomography at three different source redshifts z_s=1, 1.5, 2, explore five different smoothing scales theta_G=1, 2, 3, 5, 10 arcmin, and explicitly compare and combine the MFs with the WL power spectrum. We find that the MFs capture a substantial amount of information from non-Gaussian features of convergence maps, i.e. beyond the power spectrum. The MFs are particularly well suited to break degeneracies and to constrain the dark energy equation of state parameter w (by a factor of ~ three better than from the power spectrum alone). The non-Gaussian information derives partly from the one-point function of the convergence (through V_0, the area MF), and partly through non-linear spatial information (through combining different smoothing scales for V_0, and through V_1 and V_2, the boundary length and genus MFs, respectively). In contrast to the power spectrum, the best constraints from the MFs are obtained only when multiple smoothing scales are combined.
Weak gravitational lensing, the deflection of light by mass, is one of the best tools to constrain the growth of cosmic structure with time and reveal the nature of dark energy. I discuss the sources of systematic uncertainty in weak lensing measurements and their theoretical interpretation, including our current understanding and other options for future improvement. These include long-standing concerns such as the estimation of coherent shears from galaxy images or redshift distributions of galaxies selected based on photometric redshifts, along with systematic uncertainties that have received less attention to date because they are subdominant contributors to the error budget in current surveys. I also discuss methods for automated systematics detection using survey data of the 2020s. The goal of this review is to describe the current state of the field and what must be done so that if weak lensing measurements lead toward surprising conclusions about key questions such as the nature of dark energy, those conclusions will be credible.
Massive neutrinos influence the background evolution of the Universe as well as the growth of structure. Being able to model this effect and constrain the sum of their masses is one of the key challenges in modern cosmology. Weak-lensing cosmological constraints will also soon reach higher levels of precision with next-generation surveys like LSST, WFIRST and Euclid. We use the MassiveNus simulations to derive constraints on the sum of neutrino masses $M_{ u}$, the present-day total matter density $Omega_{rm m}$, and the primordial power spectrum normalization $A_{rm s}$ in a tomographic setting. We measure the lensing power spectrum as second-order statistics along with peak counts as higher-order statistics on lensing convergence maps generated from the simulations. We investigate the impact of multiscale filtering approaches on cosmological parameters by employing a starlet (wavelet) filter and a concatenation of Gaussian filters. In both cases peak counts perform better than the power spectrum on the set of parameters [$M_{ u}$, $Omega_{rm m}$, $A_{rm s}$] respectively by 63$%$, 40$%$ and 72$%$ when using a starlet filter and by 70$%$, 40$%$ and 77$%$ when using a multiscale Gaussian. More importantly, we show that when using a multiscale approach, joining power spectrum and peaks does not add any relevant information over considering just the peaks alone. While both multiscale filters behave similarly, we find that with the starlet filter the majority of the information in the data covariance matrix is encoded in the diagonal elements; this can be an advantage when inverting the matrix, speeding up the numerical implementation.