No Arabic abstract
Astronomy has entered the multi-messenger data era and Machine Learning has found widespread use in a large variety of applications. The exploitation of synoptic (multi-band and multi-epoch) surveys, like LSST (Legacy Survey of Space and Time), requires an extensive use of automatic methods for data processing and interpretation. With data volumes in the petabyte domain, the discrimination of time-critical information has already exceeded the capabilities of human operators and crowds of scientists have extreme difficulty to manage such amounts of data in multi-dimensional domains. This work is focused on an analysis of critical aspects related to the approach, based on Machine Learning, to variable sky sources classification, with special care to the various types of Supernovae, one of the most important subjects of Time Domain Astronomy, due to their crucial role in Cosmology. The work is based on a test campaign performed on simulated data. The classification was carried out by comparing the performances among several Machine Learning algorithms on statistical parameters extracted from the light curves. The results make in evidence some critical aspects related to the data quality and their parameter space characterization, propaedeutic to the preparation of processing machinery for the real data exploitation in the incoming decade.
The commissioning team for the Vera C. Rubin observatory is planning a set of engineering and science verification observations with the Legacy Survey of Space and Time (LSST) commissioning camera and then the Rubin Observatory LSST Camera. The time frame for these observations is not yet fixed, and the commissioning team will have flexibility in selecting fields to observe. In this document, the Dark Energy Science Collaboration (DESC) Commissioning Working Group presents a prioritized list of target fields appropriate for testing various aspects of DESC-relevant science performance, grouped by season for visibility from Rubin Observatory at Cerro Pachon. Our recommended fields include Deep-Drilling fields (DDFs) to full LSST depth for photo-$z$ and shape calibration purposes, HST imaging fields to full depth for deblending studies, and an $sim$200 square degree area to 1-year depth in several filters for higher-level validation of wide-area science cases for DESC. We also anticipate that commissioning observations will be needed for template building for transient science over a broad RA range. We include detailed descriptions of our recommended fields along with associated references. We are optimistic that this document will continue to be useful during LSST operations, as it provides a comprehensive list of overlapping data-sets and the references describing them.
The exploitation of present and future synoptic (multi-band and multi-epoch) surveys requires an extensive use of automatic methods for data processing and data interpretation. In this work, using data extracted from the Catalina Real Time Transient Survey (CRTS), we investigate the classification performance of some well tested methods: Random Forest, MLPQNA (Multi Layer Perceptron with Quasi Newton Algorithm) and K-Nearest Neighbors, paying special attention to the feature selection phase. In order to do so, several classification experiments were performed. Namely: identification of cataclysmic variables, separation between galactic and extra-galactic objects and identification of supernovae.
Strong lensing by galaxy clusters can be used to significantly expand the survey reach, thus allowing observation of magnified high-redshift supernovae that otherwise would remain undetected. Strong lensing can also provide multiple images of the galaxies that lie behind the clusters. Detection of strongly lensed Type Ia supernovae (SNe Ia) is especially useful because of their standardizable brightness, as they can be used to improve either cluster lensing models or independent measurements of cosmological parameters. The cosmological parameter, the Hubble constant, is of particular interest given the discrepancy regarding its value from measurements with different approaches. Here, we explore the feasibility of the Vera C. Rubin Observatory Legacy Survey of Space and Time (LSST) of detecting strongly lensed SNe in the field of five galaxy clusters (Abell 1689 and Hubble Frontier Fields clusters) that have well-studied lensing models. Considering the 88 systems composed of 268 individual multiple images in the five cluster fields, we find that the LSST will be sensitive to SNe~Ia (SNe~IIP) exploding in 41 (23) galaxy images. The range of redshift of these galaxies is between $1.01 < z < 3.05$. During its 10 years of operation, LSST is expected to detect $0.2pm0.1$ SN~Ia and $0.9pm0.3$ core collapse SNe. However, as LSST will observe many more massive galaxy clusters, it is likely that the expectations are higher. We stress the importance of having an additional observing program for photometric and spectroscopic follow-up of the strongly lensed SNe detected by LSST.
The advancement of technology has resulted in a rapid increase in supernova (SN) discoveries. The Subaru/Hyper Suprime-Cam (HSC) transient survey, conducted from fall 2016 through spring 2017, yielded 1824 SN candidates. This gave rise to the need for fast type classification for spectroscopic follow-up and prompted us to develop a machine learning algorithm using a deep neural network (DNN) with highway layers. This machine is trained by actual observed cadence and filter combinations such that we can directly input the observed data array into the machine without any interpretation. We tested our model with a dataset from the LSST classification challenge (Deep Drilling Field). Our classifier scores an area under the curve (AUC) of 0.996 for binary classification (SN Ia or non-SN Ia) and 95.3% accuracy for three-class classification (SN Ia, SN Ibc, or SN II). Application of our binary classification to HSC transient data yields an AUC score of 0.925. With two weeks of HSC data since the first detection, this classifier achieves 78.1% accuracy for binary classification, and the accuracy increases to 84.2% with the full dataset. This paper discusses the potential use of machine learning for SN type classification purposes.
Survey telescopes such as the Vera C. Rubin Observatory will increase the number of observed supernovae (SNe) by an order of magnitude, discovering millions of events; however, it is impossible to spectroscopically confirm the class for all the SNe discovered. Thus, photometric classification is crucial but its accuracy depends on the not-yet-finalized observing strategy of Rubin Observatorys Legacy Survey of Space and Time (LSST). We quantitatively analyze the impact of the LSST observing strategy on SNe classification using the simulated multi-band light curves from the Photometric LSST Astronomical Time-Series Classification Challenge (PLAsTiCC). First, we augment the simulated training set to be representative of the photometric redshift distribution per supernovae class, the cadence of observations, and the flux uncertainty distribution of the test set. Then we build a classifier using the photometric transient classification library snmachine, based on wavelet features obtained from Gaussian process fits, yielding similar performance to the winning PLAsTiCC entry. We study the classification performance for SNe with different properties within a single simulated observing strategy. We find that season length is an important factor, with light curves of 150 days yielding the highest classification performance. Cadence is also crucial for SNe classification; events with median inter-night gap of <3.5 days yield higher performance. Interestingly, we find that large gaps (>10 days) in light curve observations does not impact classification performance as long as sufficient observations are available on either side, due to the effectiveness of the Gaussian process interpolation. This analysis is the first exploration of the impact of observing strategy on photometric supernova classification with LSST.