Do you want to publish a course? Click here

Accurate cosmic shear errors: do we need ensembles of simulations?

63   0   0.0 ( 0 )
 Added by Alexandre Barreira
 Publication date 2018
  fields Physics
and research's language is English




Ask ChatGPT about the research

Accurate inference of cosmology from weak lensing shear requires an accurate shear power spectrum covariance matrix. Here, we investigate this accuracy requirement and quantify the relative importance of the Gaussian (G), super-sample covariance (SSC) and connected non-Gaussian (cNG) contributions to the covariance. Specifically, we forecast cosmological parameter constraints for future wide-field surveys and study how different covariance matrix components affect parameter bounds. Our main result is that the cNG term represents only a small and potentially negligible contribution to statistical parameter errors: the errors obtained using the G+SSC subset are within $lesssim 5%$ of those obtained with the full G+SSC+cNG matrix for a Euclid-like survey. This result also holds for the shear two-point correlation function, variations in survey specifications and for different analytical prescriptions of the cNG term. The cNG term is that which is often tackled using numerically expensive ensembles of survey realizations. Our results suggest however that the accuracy of analytical or approximate numerical methods to compute the cNG term is likely to be sufficient for cosmic shear inference from the next generation of surveys.



rate research

Read More

We examine the possibility of soft cosmology, namely small deviations from the usual framework due to the effective appearance of soft-matter properties in the Universe sectors. One effect of such a case would be the dark energy to exhibit a different equation-of-state parameter at large scales (which determine the universe expansion) and at intermediate scales (which determine the sub-horizon clustering and the large scale structure formation). Concerning soft dark matter, we show that it can effectively arise due to the dark-energy clustering, even if dark energy is not soft. We propose a novel parametrization introducing the softness parameters of the dark sectors. As we see, although the background evolution remains unaffected, due to the extreme sensitivity and significant effects on the global properties even a slightly non-trivial softness parameter can improve the clustering behavior and alleviate e.g. the $fsigma_8$ tension. Lastly, an extension of the cosmological perturbation theory and a detailed statistical mechanical analysis, in order to incorporate complexity and estimate the scale-dependent behavior from first principles, is necessary and would provide a robust argumentation in favour of soft cosmology.
105 - Tomohiro Nakama , Yi Wang 2018
Recently, the formation of primordial black holes (PBHs) from the collapse of primordial fluctuations has received much attention. The abundance of PBHs formed during radiation domination is sensitive to the tail of the probability distribution of primordial fluctuations. We quantify the level of fine-tuning due to this sensitivity. For example, if the main source of dark matter is PBHs with mass $10^{-12}M_odot$, then anthropic reasoning suggests that the dark matter to baryon ratio should range between 1 and 300. For this to happen, the root-mean-square amplitude of the curvature perturbation has to be fine-tuned within a $7.1%$ range. As another example, if the recently detected gravitational-wave events are to be explained by PBHs, the corresponding degree of fine-tuning is $3.8%$. We also find, however, that these fine-tunings can be relaxed if the primordial fluctuations are highly non-Gaussian, or if the PBHs are formed during an early-matter-dominated phase. We also note that no fine-tuning is needed for the scenario of a reheating of the universe by evaporated PBHs with Planck-mass relics left to serve as dark matter.
The intent recognition is an essential algorithm of any conversational AI application. It is responsible for the classification of an input message into meaningful classes. In many bot development platforms, we can configure the NLU pipeline. Several intent recognition services are currently available as an API, or we choose from many open-source alternatives. However, there is no comparison of intent recognition services and open-source algorithms. Many factors make the selection of the right approach to the intent recognition challenging in practice. In this paper, we suggest criteria to choose the best intent recognition algorithm for an application. We present a dataset for evaluation. Finally, we compare selected public NLU services with selected open-source algorithms for intent recognition.
We consider reciprocal metasurfaces with engineered reflection and transmission coefficients and study the role of normal (with respect to the metasurface plane) electric and magnetic polarizations on the possibilities to shape the reflection and transmission responses. We demonstrate in general and on a representative example that the presence of normal components of the polarization vectors does not add extra degrees of freedom in engineering the reflection and transmission characteristics of metasurfaces. Furthermore, we discuss advantages and disadvantages of equivalent volumetric and fully planar realizations of the same properties of functional metasurfaces.
The software development community has been using code quality metrics for the last five decades. Despite their wide adoption, code quality metrics have attracted a fair share of criticism. In this paper, first, we carry out a qualitative exploration by surveying software developers to gauge their opinions about current practices and potential gaps with the present set of metrics. We identify deficiencies including lack of soundness, i.e., the ability of a metric to capture a notion accurately as promised by the metric, lack of support for assessing software architecture quality, and insufficient support for assessing software testing and infrastructure. In the second part of the paper, we focus on one specific code quality metric-LCOM as a case study to explore opportunities towards improved metrics. We evaluate existing LCOM algorithms qualitatively and quantitatively to observe how closely they represent the concept of cohesion. In this pursuit, we first create eight diverse cases that any LCOM algorithm must cover and obtain their cohesion levels by a set of experienced developers and consider them as a ground truth. We show that the present set of LCOM algorithms do poorly w.r.t. these cases. To bridge the identified gap, we propose a new approach to compute LCOM and evaluate the new approach with the ground truth. We also show, using a quantitative analysis using more than 90 thousand types belonging to 261 high-quality Java repositories, the present set of methods paint a very inaccurate and misleading picture of class cohesion. We conclude that the current code quality metrics in use suffer from various deficiencies, presenting ample opportunities for the research community to address the gaps.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا