No Arabic abstract
We present a new method for generating initial conditions for numerical cosmological simulations in which massive neutrinos are treated as an extra set of N-body (collisionless) particles. It allows us to accurately follow the density field for both Cold Dark Matter (CDM) and neutrinos at both high and low redshifts. At high redshifts, the new method is able to reduce the shot noise in the neutrino power spectrum by a factor of more than $10^7$ compared to previous methods, where the power spectrum was dominated by shot noise at all scales. We find that our new approach also helps to reduce the noise on the total matter power spectrum on large scales, whereas on small scales the results agree with previous simulations. Our new method also allows for a systematic study of clustering of the low velocity tail of the distribution function of neutrinos. This method also allows for the study of the evolution of the overall velocity distribution as a function of the environment determined by the CDM field.
In the next decade, cosmological surveys will have the statistical power to detect the absolute neutrino mass scale. N-body simulations of large-scale structure formation play a central role in interpreting data from such surveys. Yet these simulations are Newtonian in nature. We provide a quantitative study of the limitations to treating neutrinos, implemented as N-body particles, in N-body codes, focusing on the error introduced by neglecting special relativistic effects. Special relativistic effects are potentially important due to the large thermal velocities of neutrino particles in the simulation box. We derive a self-consistent theory of linear perturbations in Newtonian and non-relativistic neutrinos and use this to demonstrate that N-body simulations overestimate the neutrino free-streaming scale, and cause errors in the matter power spectrum that depend on the initial redshift of the simulations. For $z_{i} lesssim 100$, and neutrino masses within the currently allowed range, this error is $lesssim 0.5%$, though represents an up to $sim 10%$ correction to the shape of the neutrino-induced suppression to the cold dark matter power spectrum. We argue that the simulations accurately model non-linear clustering of neutrinos so that the error is confined to linear scales.
Cosmological $N$-body simulations are typically purely run with particles using Newtonian equations of motion. However, such simulations can be made fully consistent with general relativity using a well-defined prescription. Here, we extend the formalism previously developed for $Lambda$CDM cosmologies with massless neutrinos to include the effects of massive, but light neutrinos. We have implemented the method in two different $N$-body codes, CONCEPT and PKDGRAV, and demonstrate that they produce consistent results. We furthermore show that we can recover all appropriate limits, including the full GR solution in linear perturbation theory at the per mille level of precision.
We demonstrate that testing for self-similarity in scale-free simulations provides an excellent tool to quantify the resolution at small scales of cosmological N-body simulations. Analysing two-point correlation functions measured in simulations using ABACUS, we show how observed deviations from self-similarity reveal the range of time and distance scales in which convergence is obtained. While the well-converged scales show accuracy below 1 percent, our results show that, with a small force softening length, the spatial resolution is essentially determined by the mass resolution. At later times the lower cut-off scale on convergence evolves in comoving units as $a^{-1/2}$ ($a$ being the scale factor), consistent with a hypothesis that it is set by two-body collisionality. A corollary of our results is that N-body simulations, particularly at high red-shift, contain a significant spatial range in which clustering appears converged with respect to the time-stepping and force softening but has not actually converged to the physical continuum result. The method developed can be applied to determine the resolution of any clustering statistic and extended to infer resolution limits for non-scale-free simulations.
Cosmology is entering an era of percent level precision due to current large observational surveys. This precision in observation is now demanding more accuracy from numerical methods and cosmological simulations. In this paper, we study the accuracy of $N$-body numerical simulations and their dependence on changes in the initial conditions and in the simulation algorithms. For this purpose, we use a series of cosmological $N$-body simulations with varying initial conditions. We test the influence of the initial conditions, namely the pre-initial configuration (preIC), the order of the Lagrangian perturbation theory (LPT), and the initial redshift, on the statistics associated with the large scale structures of the universe such as the halo mass function, the density power spectrum, and the maximal extent of the large scale structures. We find that glass or grid pre-initial conditions give similar results at $zlesssim 2$. However, the initial excess of power in the glass initial conditions yields a subtle difference in the power spectra and the mass function at high redshifts. The LPT order used to generate the ICs of the simulations is found to play a crucial role. First-order LPT (1LPT) simulations underestimate the number of massive haloes with respect to second-order (2LPT) ones, typically by 2% at $10^{14} h^{-1} M_odot$ for an initial redshift of 23, and the small-scale power with an underestimation of 6% near the Nyquist frequency for $z_mathrm{ini} = 23$. Moreover, at higher redshifts, the high-mass end of the mass function is significantly underestimated in 1LPT simulations. On the other hand, when the LPT order is fixed, the starting redshift has a systematic impact on the low-mass end of the halo mass function.
Gravitational softening length is one of the key parameters to properly set up a cosmological $N$-body simulation. In this paper, we perform a large suit of high-resolution $N$-body simulations to revise the optimal softening scheme proposed by Power et al. (P03). Our finding is that P03 optimal scheme works well but is over conservative. Using smaller softening lengths than that of P03 can achieve higher spatial resolution and numerically convergent results on both circular velocity and density profiles. However using an over small softening length overpredicts matter density at the inner most region of dark matter haloes. We empirically explore a better optimal softening scheme based on P03 form and find that a small modification works well. This work will be useful for setting up cosmological simulations.