ترغب بنشر مسار تعليمي؟ اضغط هنا

DataVault: A Data Storage Infrastructure for the Einstein Toolkit

308   0   0.0 ( 0 )
 نشر من قبل Yufeng Luo
 تاريخ النشر 2020
والبحث باللغة English




اسأل ChatGPT حول البحث

Data sharing is essential in the numerical simulations research. We introduce a data repository, DataVault, that is designed for data sharing, search and analysis. A comparative study of existing repositories is performed to analyze features that are critical to a data repository. We describe the architecture, workflow, and deployment of DataVault, and provide three use-case scenarios for different communities to facilitate the use and application of DataVault. Potential features are proposed and we outline the future development for these features.

قيم البحث

اقرأ أيضاً

Numerical relativity codes that do not make assumptions on spatial symmetries most commonly adopt Cartesian coordinates. While these coordinates have many attractive features, spherical coordinates are much better suited to take advantage of approxim ate symmetries in a number of astrophysical objects, including single stars, black holes and accretion disks. While the appearance of coordinate singularities often spoils numerical relativity simulations in spherical coordinates, especially in the absence of any symmetry assumptions, it has recently been demonstrated that these problems can be avoided if the coordinate singularities are handled analytically. This is possible with the help of a reference-metric version of the Baumgarte-Shapiro-Shibata-Nakamura formulation together with a proper rescaling of tensorial quantities. In this paper we report on an implementation of this formalism in the Einstein Toolkit. We adapt the Einstein Toolkit infrastructure, originally designed for Cartesian coordinates, to handle spherical coordinates, by providing appropriate boundary conditions at both inner and outer boundaries. We perform numerical simulations for a disturbed Kerr black hole, extract the gravitational wave signal, and demonstrate that the noise in these signals is orders of magnitude smaller when computed on spherical grids rather than Cartesian grids. With the public release of our new Einstein Toolkit thorns, our methods for numerical relativity in spherical coordinates will become available to the entire numerical relativity community.
180 - Yupeng Fu , Chinmay Soman 2021
Ubers business is highly real-time in nature. PBs of data is continuously being collected from the end users such as Uber drivers, riders, restaurants, eaters and so on everyday. There is a lot of valuable information to be processed and many decisio ns must be made in seconds for a variety of use cases such as customer incentives, fraud detection, machine learning model prediction. In addition, there is an increasing need to expose this ability to different user categories, including engineers, data scientists, executives and operations personnel which adds to the complexity. In this paper, we present the overall architecture of the real-time data infrastructure and identify three scaling challenges that we need to continuously address for each component in the architecture. At Uber, we heavily rely on open source technologies for the key areas of the infrastructure. On top of those open-source software, we add significant improvements and customizations to make the open-source solutions fit in Ubers environment and bridge the gaps to meet Ubers unique scale and requirements. We then highlight several important use cases and show their real-time solutions and tradeoffs. Finally, we reflect on the lessons we learned as we built, operated and scaled these systems.
Einstein Telescope (ET) is conceived to be a third generation gravitational-wave observatory. Its amplitude sensitivity would be a factor ten better than advanced LIGO and Virgo and it could also extend the low-frequency sensitivity down to 1--3 Hz, compared to the 10--20 Hz of advanced detectors. Such an observatory will have the potential to observe a variety of different GW sources, including compact binary systems at cosmological distances. ETs expected reach for binary neutron star (BNS) coalescences is out to redshift $zsimeq 2$ and the rate of detectable BNS coalescences could be as high as one every few tens or hundreds of seconds, each lasting up to several days. %in the sensitive frequency band of ET. With such a signal-rich environment, a key question in data analysis is whether overlapping signals can be discriminated. In this paper we simulate the GW signals from a cosmological population of BNS and ask the following questions: Does this population create a confusion background that limits ETs ability to detect foreground sources? How efficient are current algorithms in discriminating overlapping BNS signals? Is it possible to discern the presence of a population of signals in the data by cross-correlating data from different detectors in the ET observatory? We find that algorithms currently used to analyze LIGO and Virgo data are already powerful enough to detect the sources expected in ET, but new algorithms are required to fully exploit ET data.
We document the data transfer workflow, data transfer performance, and other aspects of staging approximately 56 terabytes of climate model output data from the distributed Coupled Model Intercomparison Project (CMIP5) archive to the National Energy Research Supercomputing Center (NERSC) at the Lawrence Berkeley National Laboratory required for tracking and characterizing extratropical storms, a phenomena of importance in the mid-latitudes. We present this analysis to illustrate the current challenges in assembling multi-model data sets at major computing facilities for large-scale studies of CMIP5 data. Because of the larger archive size of the upcoming CMIP6 phase of model intercomparison, we expect such data transfers to become of increasing importance, and perhaps of routine necessity. We find that data transfer rates using the ESGF are often slower than what is typically available to US residences and that there is significant room for improvement in the data transfer capabilities of the ESGF portal and data centers both in terms of workflow mechanics and in data transfer performance. We believe performance improvements of at least an order of magnitude are within technical reach using current best practices, as illustrated by the performance we achieved in transferring the complete raw data set between two high performance computing facilities. To achieve these performance improvements, we recommend: that current best practices (such as the Science DMZ model) be applied to the data servers and networks at ESGF data centers; that sufficient financial and human resources be devoted at the ESGF data centers for systems and network engineering tasks to support high performance data movement; and that performance metrics for data transfer between ESGF data centers and major computing facilities used for climate data analysis be established, regularly tested, and published.
We present SphericalNR, a new framework for the publicly available Einstein Toolkit that numerically solves the Einstein field equations coupled to the equations of general relativistic magnetohydrodynamics (GRMHD) in a 3+1 split of spacetime in sphe rical coordinates without symmetry assumptions. The spacetime evolution is performed using reference-metr
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا