No Arabic abstract
The two-point correlation function of the galaxy distribution is a key cosmological observable that allows us to constrain the dynamical and geometrical state of our Universe. To measure the correlation function we need to know both the galaxy positions and the expected galaxy density field. The expected field is commonly specified using a Monte-Carlo sampling of the volume covered by the survey and, to minimize additional sampling errors, this random catalog has to be much larger than the data catalog. Correlation function estimators compare data-data pair counts to data-random and random-random pair counts, where random-random pairs usually dominate the computational cost. Future redshift surveys will deliver spectroscopic catalogs of tens of millions of galaxies. Given the large number of random objects required to guarantee sub-percent accuracy, it is of paramount importance to improve the efficiency of the algorithm without degrading its precision. We show both analytically and numerically that splitting the random catalog into a number of subcatalogs of the same size as the data catalog when calculating random-random pairs, and excluding pairs across different subcatalogs provides the optimal error at fixed computational cost. For a random catalog fifty times larger than the data catalog, this reduces the computation time by a factor of more than ten without affecting estimator variance or bias.
We perform theoretical and numerical studies of the full relativistic two-point galaxy correlation function, considering the linear-order scalar and tensor perturbation contributions and the wide-angle effects. Using the gauge-invariant relativistic description of galaxy clustering and accounting for the contributions at the observer position, we demonstrate that the complete theoretical expression is devoid of any long-mode contributions from scalar or tensor perturbations and it lacks the infrared divergences in agreement with the equivalence principle. By showing that the gravitational potential contribution to the correlation function converges in the infrared, our study justifies an IR cut-off $(k_{text{IR}} leq H_0)$ in computing the gravitational potential contribution. Using the full gauge-invariant expression, we numerically compute the galaxy two-point correlation function and study the individual contributions in the conformal Newtonian gauge. We find that the terms at the observer position such as the coordinate lapses and the observer velocity (missing in the standard formalism) dominate over the other relativistic contributions in the conformal Newtonian gauge such as the source velocity, the gravitational potential, the integrated Sachs-Wolf effect, the Shapiro time-delay and the lensing convergence. Compared to the standard Newtonian theoretical predictions that consider only the density fluctuation and redshift-space distortions, the relativistic effects in galaxy clustering result in a few percent-level systematic errors beyond the scale of the baryonic acoustic oscillation. Our theoretical and numerical study provides a comprehensive understanding of the relativistic effects in the galaxy two-point correlation function, as it proves the validity of the theoretical prediction and accounts for effects that are often neglected in its numerical evaluation.
All estimators of the two-point correlation function are based on a random catalogue, a set of points with no intrinsic clustering following the selection function of a survey. High-accuracy estimates require the use of large random catalogues, which imply a high computational cost. We propose to replace the standard random catalogues by glass-like point distributions or glass catalogues, which are characterized by a power spectrum $P(k)propto k^4$ and exhibit significantly less power than a Poisson distribution with the same number of points on scales larger than the mean inter-particle separation. We show that these distributions can be obtained by iteratively applying the technique of Zeldovich reconstruction commonly used in studies of baryon acoustic oscillations (BAO). We provide a modified version of the widely used Landy-Szalay estimator of the correlation function adapted to the use of glass catalogues and compare its performance with the results obtained using random samples. Our results show that glass-like samples do not add any bias with respect to the results obtained using Poisson distributions. On scales larger than the mean inter-particle separation of the glass catalogues, the modified estimator leads to a significant reduction of the variance of the Legendre multipoles $xi_ell(s)$ with respect to the standard Landy-Szalay results with the same number of points. The size of the glass catalogue required to achieve a given accuracy in the correlation function is significantly smaller than when using random samples. Even considering the small additional cost of constructing the glass catalogues, their use could help to drastically reduce the computational cost of configuration-space clustering analysis of future surveys while maintaining high-accuracy requirements.
We investigate how the shape of the galaxy two-point correlation function as measured in the zCOSMOS survey depends on local environment, quantified in terms of the density contrast on scales of 5 Mpc/h. We show that the flat shape previously observed at redshifts between z=0.6 and z=1 can be explained by this volume being simply 10% over-abundant in high-density environments, with respect to a Universal density probability distribution function. When galaxies corresponding to the top 10% tail of the distribution are excluded, the measured w_p(r_p) steepens and becomes indistinguishable from LCDM predictions on all scales. This is the same effect recognised by Abbas & Sheth in the SDSS data at z~0 and explained as a natural consequence of halo-environment correlations in a hierarchical scenario. Galaxies living in high-density regions trace dark matter halos with typically higher masses, which are more correlated. If the density probability distribution function of the sample is particularly rich in high-density regions because of the variance introduced by its finite size, this produces a distorted two-point correlation function. We argue that this is the dominant effect responsible for the observed peculiar clustering in the COSMOS field.
The growth history of large-scale structure in the Universe is a powerful probe of the cosmological model, including the nature of dark energy. We study the growth rate of cosmic structure to redshift $z = 0.9$ using more than $162{,}000$ galaxy redshifts from the WiggleZ Dark Energy Survey. We divide the data into four redshift slices with effective redshifts $z = [0.2,0.4,0.6,0.76]$ and in each of the samples measure and model the 2-point galaxy correlation function in parallel and transverse directions to the line-of-sight. After simultaneously fitting for the galaxy bias factor we recover values for the cosmic growth rate which are consistent with our assumed $Lambda$CDM input cosmological model, with an accuracy of around 20% in each redshift slice. We investigate the sensitivity of our results to the details of the assumed model and the range of physical scales fitted, making close comparison with a set of N-body simulations for calibration. Our measurements are consistent with an independent power-spectrum analysis of a similar dataset, demonstrating that the results are not driven by systematic errors. We determine the pairwise velocity dispersion of the sample in a non-parametric manner, showing that it systematically increases with decreasing redshift, and investigate the Alcock-Paczynski effects of changing the assumed fiducial model on the results. Our techniques should prove useful for current and future galaxy surveys mapping the growth rate of structure using the 2-dimensional correlation function.
The increasingly large amount of cosmological data coming from ground-based and space-borne telescopes requires highly efficient and fast enough data analysis techniques to maximise the scientific exploitation. In this work, we explore the capabilities of supervised machine learning algorithms to learn the properties of the large-scale structure of the Universe, aiming at constraining the matter density parameter, Omega m. We implement a new Artificial Neural Network for a regression data analysis, and train it on a large set of galaxy two-point correlation functions in standard cosmologies with different values of Omega m. The training set is constructed from log-normal mock catalogues which reproduce the clustering of the Baryon Oscillation Spectroscopic Survey (BOSS) galaxies. The presented statistical method requires no specific analytical model to construct the likelihood function, and runs with negligible computational cost, after training. We test this new Artificial Neural Network on real BOSS data, finding Omega m=0.309p/m0.008, which is remarkably consistent with standard analysis results.