No Arabic abstract
Establishing accurate morphological measurements of galaxies in a reasonable amount of time for future big-data surveys such as EUCLID, the Large Synoptic Survey Telescope or the Wide Field Infrared Survey Telescope is a challenge. Because of its high level of abstraction with little human intervention, deep learning appears to be a promising approach. Deep learning is a rapidly growing discipline that models high-level patterns in data as complex multilayered networks. In this work we test the ability of deep convolutional networks to provide parametric properties of Hubble Space Telescope like galaxies (half-light radii, Sersic indices, total flux etc..). We simulate a set of galaxies including point spread function and realistic noise from the CANDELS survey and try to recover the main galaxy parameters using deep-learning. We com- pare the results with the ones obtained with the commonly used profile fitting based software GALFIT. This way showing that with our method we obtain results at least equally good as the ones obtained with GALFIT but, once trained, with a factor 5 hundred time faster.
Understanding morphological types of galaxies is a key parameter for studying their formation and evolution. Neural networks that have been used previously for galaxy morphology classification have some disadvantages, such as not being invariant under rotation. In this work, we studied the performance of Capsule Network, a recently introduced neural network architecture that is rotationally invariant and spatially aware, on the task of galaxy morphology classification. We designed two evaluation scenarios based on the answers from the question tree in the Galaxy Zoo project. In the first scenario, we used Capsule Network for regression and predicted probabilities for all of the questions. In the second scenario, we chose the answer to the first morphology question that had the highest user agreement as the class of the object and trained a Capsule Network classifier, where we also reconstructed galaxy images. We achieved promising results in both of these scenarios. Automated approaches such as the one introduced here will greatly decrease the workload of astronomers and will play a critical role in the upcoming large sky surveys.
We use a continuous depth version of the Residual Network (ResNet) model known as Neural ordinary differential equations (NODE) for the purpose of galaxy morphology classification. We applied this method to carry out supervised classification of galaxy images from the Galaxy Zoo 2 dataset, into five distinct classes, and obtained an accuracy of about 92% for most of the classes. Through our experiments, we show that NODE not only performs as well as other deep neural networks, but has additional advantages over them, which can prove very useful for next generation surveys. We also compare our result against ResNet. While ResNet and its variants suffer problems, such as time consuming architecture selection (e.g. the number of layers) and the requirement of large data for training, NODE does not have these requirements. Through various metrics, we conclude that the performance of NODE matches that of other models, despite using only one-third of the total number of parameters as compared to these other models.
The morphological classification of galaxies is a relevant probe for galaxy evolution and unveils its connection with cosmological structure formation. To this scope, it is fundamental to recover galaxy morphologies over large areas of the sky. In this paper, we present a morphological catalogue for galaxies in the Stripe-82 area, observed with S-PLUS, till a magnitude limit of $rle17$, using the state-of-the-art of Convolutional Neural Networks (CNNs) for computer vision. This analysis will then be extended to the whole S-PLUS survey data, covering $simeq 9300$ $deg^{2}$ of the celestial sphere in twelve optical bands. We find that the networks performance increases with 5 broad bands and additional 3 narrow bands compared to our baseline with 3 bands. However, it does lose performance when using the full $12$ band image information. Nevertheless, the best result is achieved with 3 bands, when using pre-trained network weights in an ImageNet dataset. These results enhance the importance of previous knowledge in the neural network weights based on training in non related extensive datasets. Thus, we release a model pre-trained in several bands that could be adapted to other surveys. We develop a catalogue of 3274 galaxies in Stripe-82 that are not presented in Galaxy Zoo 1 (GZ1). We also add classification to 4686 galaxies considered ambiguous in GZ1 dataset. Finally, we present a prospect of a novel way to take advantage of $12$ bands information for morphological classification using multiband morphometric features. The morphological catalogues are publicly available.
With the increasing number of deep multi-wavelength galaxy surveys, the spectral energy distribution (SED) of galaxies has become an invaluable tool for studying the formation of their structures and their evolution. In this context, standard analysis relies on simple spectro-photometric selection criteria based on a few SED colors. If this fully supervised classification already yielded clear achievements, it is not optimal to extract relevant information from the data. In this article, we propose to employ very recent advances in machine learning, and more precisely in feature learning, to derive a data-driven diagram. We show that the proposed approach based on denoising autoencoders recovers the bi-modality in the galaxy population in an unsupervised manner, without using any prior knowledge on galaxy SED classification. This technique has been compared to principal component analysis (PCA) and to standard color/color representations. In addition, preliminary results illustrate that this enables the capturing of extra physically meaningful information, such as redshift dependence, galaxy mass evolution and variation over the specific star formation rate. PCA also results in an unsupervised representation with physical properties, such as mass and sSFR, although this representation separates out. less other characteristics (bimodality, redshift evolution) than denoising autoencoders.
Deconvolution of large survey images with millions of galaxies requires to develop a new generation of methods which can take into account a space variant Point Spread Function (PSF) and have to be at the same time accurate and fast. We investigate in this paper how Deep Learning (DL) could be used to perform this task. We employ a U-Net Deep Neural Network (DNN) architecture to learn in a supervised setting parameters adapted for galaxy image processing and study two strategies for deconvolution. The first approach is a post-processing of a mere Tikhonov deconvolution with closed form solution and the second one is an iterative deconvolution framework based on the Alternating Direction Method of Multipliers (ADMM). Our numerical results based on GREAT3 simulations with realistic galaxy images and PSFs show that our two approaches outperforms standard techniques based on convex optimization, whether assessed in galaxy image reconstruction or shape recovery. The approach based on Tikhonov deconvolution leads to the most accurate results except for ellipticity errors at high signal to noise ratio where the ADMM approach performs slightly better, is also more computation-time efficient to process a large number of galaxies, and is therefore recommended in this scenario.