No Arabic abstract
N-body simulations are widely used to simulate the dynamical evolution of a variety of systems, among them star clusters. Much of our understanding of their evolution rests on the results of such direct N-body simulations. They provide insight in the structural evolution of star clusters, as well as into the occurrence of stellar exotica. Although the major pure N-body codes STARLAB/KIRA and NBODY4 are widely used for a range of applications, there is no thorough comparison study yet. Here we thoroughly compare basic quantities as derived from simulations performed either with STARLAB/KIRA or NBODY4. We construct a large number of star cluster models for various stellar mass function settings (but without stellar/binary evolution, primordial binaries, external tidal fields etc), evolve them in parallel with STARLAB/KIRA and NBODY4, analyse them in a consistent way and compare the averaged results quantitatively. For this quantitative comparison we develop a bootstrap algorithm for functional dependencies. We find an overall excellent agreement between the codes, both for the clusters structural and energy parameters as well as for the properties of the dynamically created binaries. However, we identify small differences, like in the energy conservation before core collapse and the energies of escaping stars, which deserve further studies. Our results reassure the comparability and the possibility to combine results from these two major N-body codes, at least for the purely dynamical models (i.e. without stellar/binary evolution) we performed. (abridged)
Most recent progress in understanding the dynamical evolution of star clusters relies on direct N-body simulations. Owing to the computational demands, and the desire to model more complex and more massive star clusters, hardware calculational accelerators, such as GRAPE special-purpose hardware or, more recently, GPUs (i.e. graphics cards), are generally utilised. In addition, simulations can be accelerated by adjusting parameters determining the calculation accuracy (i.e. changing the internal simulation time step used for each star). We extend our previous thorough comparison (Anders et al. 2009) of basic quantities as derived from simulations performed either with STARLAB/KIRA or NBODY6. Here we focus on differences arising from using different hardware accelerations (including the increasingly popular graphic card accelerations/GPUs) and different calculation accuracy settings. We use the large number of star cluster models (for a fixed stellar mass function, without stellar/binary evolution, primordial binaries, external tidal fields etc) already used in the previous paper, evolve them with STARLAB/KIRA (and NBODY6, where required), analyse them in a consistent way and compare the averaged results quantitatively. For this quantitative comparison, we apply the bootstrap algorithm for functional dependencies developed in our previous study. In general we find very high comparability of the simulation results, independent of the used computer hardware (including the hardware accelerators) and the used N-body code. For the tested accuracy settings we find that for reduced accuracy (i.e. time step at least a factor 2.5 larger than the standard setting) most simulation results deviate significantly from the results using standard settings. The remaining deviations are comprehensible and explicable.
We give an overview about equations of state (EOS) which are currently available for simulations of core-collapse supernovae and neutron star mergers. A few selected important aspects of the EOS, such as the symmetry energy, the maximum mass of neutron stars, and cluster formation, are confronted with constraints from experiments and astrophysical observations. There are just very few models which are compatible even with this very restricted set of constraints. These remaining models illustrate the uncertainty of the uniform nuclear matter EOS at high densities. In addition, at finite temperatures the medium modifications of nuclear clusters represent a conceptual challenge. In conclusion, there has been significant development in the recent years, but there is still need for further improved general purpose EOS tables.
Cosmological simulations of galaxy formation often rely on prescriptions for star formation and feedback that depend on halo properties such as halo mass, central over-density, and virial temperature. In this paper we address the convergence of individual halo properties, based on their number of particles N, focusing in particular on the mass of halos near the resolution limit of a simulation. While it has been established that the halo mass function is sampled on average down to N~30 particles, we show that individual halo properties exhibit significant scatter, and some systematic biases, as one approaches the resolution limit. We carry out a series of cosmological simulations using the Gadget2 and Enzo codes with N_p=64^3 to N_p=1024^3 total particles, keeping the same large-scale structure in the simulation box. We consider boxes from l_{box} = 8 Mpc/h to l_{box} = 512 Mpc/h to probe different halo masses and formation redshifts. We cross-identify dark matter halos in boxes at different resolutions and measure the scatter in their properties. The uncertainty in the mass of single halos depends on the number of particles (scaling approximately as N^{-1/3}), but the rarer the density peak, the more robust its identification. The virial radius of halos is very stable and can be measured without bias for halos with N>30. In contrast, the average density within a sphere containing 25% of the total halo mass is severely underestimated (by more than a factor 2) and the halo spin is moderately overestimated for N<100. If sub-grid physics is implemented upon a cosmological simulation, we recommend that rare halos (~3sigma peaks) be resolved with N>100 particles and common halos (~1sigma peaks) with N>400 particles to avoid excessive numerical noise and possible systematic biases in the results.
The graph of a Bayesian Network (BN) can be machine learned, determined by causal knowledge, or a combination of both. In disciplines like bioinformatics, applying BN structure learning algorithms can reveal new insights that would otherwise remain unknown. However, these algorithms are less effective when the input data are limited in terms of sample size, which is often the case when working with real data. This paper focuses on purely machine learned and purely knowledge-based BNs and investigates their differences in terms of graphical structure and how well the implied statistical models explain the data. The tests are based on four previous case studies whose BN structure was determined by domain knowledge. Using various metrics, we compare the knowledge-based graphs to the machine learned graphs generated from various algorithms implemented in TETRAD spanning all three classes of learning. The results show that, while the algorithms produce graphs with much higher model selection score, the knowledge-based graphs are more accurate predictors of variables of interest. Maximising score fitting is ineffective in the presence of limited sample size because the fitting becomes increasingly distorted with limited data, guiding algorithms towards graphical patterns that share higher fitting scores and yet deviate considerably from the true graph. This highlights the value of causal knowledge in these cases, as well as the need for more appropriate fitting scores suitable for limited data. Lastly, the experiments also provide new evidence that support the notion that results from simulated data tell us little about actual real-world performance.
We describe Space Warps, a novel gravitational lens discovery service that yields samples of high purity and completeness through crowd-sourced visual inspection. Carefully produced colour composite images are displayed to volunteers via a web- based classification interface, which records their estimates of the positions of candidate lensed features. Images of simulated lenses, as well as real images which lack lenses, are inserted into the image stream at random intervals; this training set is used to give the volunteers instantaneous feedback on their performance, as well as to calibrate a model of the system that provides dynamical updates to the probability that a classified image contains a lens. Low probability systems are retired from the site periodically, concentrating the sample towards a set of lens candidates. Having divided 160 square degrees of Canada-France-Hawaii Telescope Legacy Survey (CFHTLS) imaging into some 430,000 overlapping 82 by 82 arcsecond tiles and displaying them on the site, we were joined by around 37,000 volunteers who contributed 11 million image classifications over the course of 8 months. This Stage 1 search reduced the sample to 3381 images containing candidates; these were then refined in Stage 2 to yield a sample that we expect to be over 90% complete and 30% pure, based on our analysis of the volunteers performance on training images. We comment on the scalability of the SpaceWarps system to the wide field survey era, based on our projection that searches of 10$^5$ images could be performed by a crowd of 10$^5$ volunteers in 6 days.