Do you want to publish a course? Click here

Science-Driven Optimization of the LSST Observing Strategy

295   0   0.0 ( 0 )
 Added by Phil Marshall
 Publication date 2017
  fields Physics
and research's language is English




Ask ChatGPT about the research

The Large Synoptic Survey Telescope is designed to provide an unprecedented optical imaging dataset that will support investigations of our Solar System, Galaxy and Universe, across half the sky and over ten years of repeated observation. However, exactly how the LSST observations will be taken (the observing strategy or cadence) is not yet finalized. In this dynamically-evolving community white paper, we explore how the detailed performance of the anticipated science investigations is expected to depend on small changes to the LSST observing strategy. Using realistic simulations of the LSST schedule and observation properties, we design and compute diagnostic metrics and Figures of Merit that provide quantitative evaluations of different observing strategies, analyzing their impact on a wide range of proposed science projects. This is work in progress: we are using this white paper to communicate to each other the relative merits of the observing strategy choices that could be made, in an effort to maximize the scientific value of the survey. The investigation of some science cases leads to suggestions for new strategies that could be simulated and potentially adopted. Notably, we find motivation for exploring departures from a spatially uniform annual tiling of the sky: focusing instead on different parts of the survey area in different years in a rolling cadence is likely to have significant benefits for a number of time domain and moving object astronomy projects. The communal assembly of a suite of quantified and homogeneously coded metrics is the vital first step towards an automated, systematic, science-based assessment of any given cadence simulation, that will enable the scheduling of the LSST to be as well-informed as possible.



rate research

Read More

A survey that can cover the sky in optical bands over wide fields to faint magnitudes with a fast cadence will enable many of the exciting science opportunities of the next decade. The Large Synoptic Survey Telescope (LSST) will have an effective aperture of 6.7 meters and an imaging camera with field of view of 9.6 deg^2, and will be devoted to a ten-year imaging survey over 20,000 deg^2 south of +15 deg. Each pointing will be imaged 2000 times with fifteen second exposures in six broad bands from 0.35 to 1.1 microns, to a total point-source depth of r~27.5. The LSST Science Book describes the basic parameters of the LSST hardware, software, and observing plans. The book discusses educational and outreach opportunities, then goes on to describe a broad range of science that LSST will revolutionize: mapping the inner and outer Solar System, stellar populations in the Milky Way and nearby galaxies, the structure of the Milky Way disk and halo and other objects in the Local Volume, transient and variable objects both at low and high redshift, and the properties of normal and active galaxies at low and high redshift. It then turns to far-field cosmological topics, exploring properties of supernovae to z~1, strong and weak lensing, the large-scale distribution of galaxies and baryon oscillations, and how these different probes may be combined to constrain cosmological models and the physics of dark energy.
The Wide-Field Infrared Survey Telescope (WFIRST) is expected to launch in the mid-2020s. With its wide-field near-infrared (NIR) camera, it will survey the sky to unprecedented detail. As part of normal operations and as the result of multiple expected dedicated surveys, WFIRST will produce several relatively wide-field (tens of square degrees) deep (limiting magnitude of 28 or fainter) fields. In particular, a planned supernova survey is expected to image 3 deep fields in the LSST footprint roughly every 5 days over 2 years. Stacking all data, this survey will produce, over all WFIRST supernova fields in the LSST footprint, ~12-25 deg^2 and ~5-15 deg^2 regions to depths of ~28 mag and ~29 mag, respectively. We suggest LSST undertake mini-surveys that will match the WFIRST cadence and simultaneously observe the supernova survey fields during the 2-year WFIRST supernova survey, achieving a stacked depth similar to that of the WFIRST data. We also suggest additional observations of these same regions throughout the LSST survey to get deep images earlier, have long-term monitoring in the fields, and produce deeper images overall. These fields will provide a legacy for cosmology, extragalactic, and transient/variable science.
The LSST survey will provide unprecedented statistical power for measurements of dark energy. Consequently, controlling systematic uncertainties is becoming more important than ever. The LSST observing strategy will affect the statistical uncertainty and systematics control for many science cases; here, we focus on weak lensing systematics. The fact that the LSST observing strategy involves hundreds of visits to the same sky area provides new opportunities for systematics mitigation. We explore these opportunities by testing how different dithering strategies (pointing offsets and rotational angle of the camera in different exposures) affect additive weak lensing shear systematics on a baseline operational simulation, using the $rho-$statistics formalism. Some dithering strategies improve systematics control at the end of the survey by a factor of up to $sim 3-4$ better than others. We find that a random translational dithering strategy, applied with random rotational dithering at every filter change, is the most effective of those strategies tested in this work at averaging down systematics. Adopting this dithering algorithm, we explore the effect of varying the area of the survey footprint, exposure time, number of exposures in a visit, and exposure to the Galactic plane. We find that any change that increases the average number of exposures (in filters relevant to weak lensing) reduces the additive shear systematics. Some ways to achieve this increase may not be favorable for the weak lensing statistical constraining power or for other probes, and we explore the relative trade-offs between these options given constraints on the overall survey parameters.
Cosmology is one of the four science pillars of LSST, which promises to be transformative for our understanding of dark energy and dark matter. The LSST Dark Energy Science Collaboration (DESC) has been tasked with deriving constraints on cosmological parameters from LSST data. Each of the cosmological probes for LSST is heavily impacted by the choice of observing strategy. This white paper is written by the LSST DESC Observing Strategy Task Force (OSTF), which represents the entire collaboration, and aims to make recommendations on observing strategy that will benefit all cosmological analyses with LSST. It is accompanied by the DESC DDF (Deep Drilling Fields) white paper (Scolnic et al.). We use a variety of metrics to understand the effects of the observing strategy on measurements of weak lensing, large-scale structure, clusters, photometric redshifts, supernovae, strong lensing and kilonovae. In order to reduce systematic uncertainties, we conclude that the current baseline observing strategy needs to be significantly modified to result in the best possible cosmological constraints. We provide some key recommendations: moving the WFD (Wide-Fast-Deep) footprint to avoid regions of high extinction, taking visit pairs in different filters, changing the 2x15s snaps to a single exposure to improve efficiency, focusing on strategies that reduce long gaps (>15 days) between observations, and prioritizing spatial uniformity at several intervals during the 10-year survey.
The observing strategy of a galaxy survey influences the degree to which its resulting data can be used to accomplish any science goal. LSST is thus seeking metrics of observing strategies for multiple science cases in order to optimally choose a cadence. Photometric redshifts are essential for many extragalactic science applications of LSSTs data, including but not limited to cosmology, but there are few metrics available, and they are not straightforwardly integrated with metrics of other cadence-dependent quantities that may influence any given use case. We propose a metric for observing strategy optimization based on the potentially recoverable mutual information about redshift from a photometric sample under the constraints of a realistic observing strategy. We demonstrate a tractable estimation of a variational lower bound of this mutual information implemented in a public code using conditional normalizing flows. By comparing the recoverable redshift information across observing strategies, we can distinguish between those that preclude robust redshift constraints and those whose data will preserve more redshift information, to be generically utilized in a downstream analysis. We recommend the use of this versatile metric to observing strategy optimization for redshift-dependent extragalactic use cases, including but not limited to cosmology, as well as any other science applications for which photometry may be modeled from true parameter values beyond redshift.
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا