No Arabic abstract
Metacalibration is a new technique for measuring weak gravitational lensing shear that is unbiased for isolated galaxy images. In this work we test metacalibration with overlapping, or ``blended galaxy images. Using standard metacalibration, we find a few percent shear measurement bias for galaxy densities relevant for current surveys, and that this bias increases with increasing galaxy number density. We show that this bias is not due to blending itself, but rather to shear-dependent object detection. If object detection is shear independent, no deblending of images is needed, in principle. We demonstrate that detection biases are accurately removed when including object detection in the metacalibration process, a technique we call metadetection. This process involves applying an artificial shear to images of small regions of sky and performing detection on the sheared images, as well as measurements that are used to calculate a shear response. We demonstrate that the method can accurately recover weak shear signals even in highly blended scenes. In the metacalibration process, the space between objects is sheared coherently, which does not perfectly match the real universe in which some, but not all, galaxy images are sheared coherently. We find that even for the worst case scenario, in which the space between objects is completely unsheared, the resulting shear bias is at most a few tenths of a percent for future surveys. We discuss additional technical challenges that must be met in order to implement metadetection for real surveys.
Metacalibration is a recently introduced method to accurately measure weak gravitational lensing shear using only the available imaging data, without need for prior information about galaxy properties or calibration from simulations. The method involves distorting the image with a small known shear, and calculating the response of a shear estimator to that applied shear. The method was shown to be accurate in moderate sized simulations with galaxy images that had relatively high signal-to-noise ratios, and without significant selection effects. In this work we introduce a formalism to correct for both shear response and selection biases. We also observe that, for images with relatively low signal-to-noise ratios, the correlated noise that arises during the metacalibration process results in significant bias, for which we develop a simple empirical correction. To test this formalism, we created large image simulations based on both parametric models and real galaxy images, including tests with realistic point-spread functions. We varied the point-spread function ellipticity at the five percent level. In each simulation we applied a small, few percent shear to the galaxy images. We introduced additional challenges that arise in real data, such as detection thresholds, stellar contamination, and missing data. We applied cuts on the measured galaxy properties to induce significant selection effects. Using our formalism, we recovered the input shear with an accuracy better than a part in a thousand in all cases.
Machine learning is a tool for building models that accurately represent input training data. When undesired biases concerning demographic groups are in the training data, well-trained models will reflect those biases. We present a framework for mitigating such biases by including a variable for the group of interest and simultaneously learning a predictor and an adversary. The input to the network X, here text or census data, produces a prediction Y, such as an analogy completion or income bracket, while the adversary tries to model a protected variable Z, here gender or zip code. The objective is to maximize the predictors ability to predict Y while minimizing the adversarys ability to predict Z. Applied to analogy completion, this method results in accurate predictions that exhibit less evidence of stereotyping Z. When applied to a classification task using the UCI Adult (Census) Dataset, it results in a predictive model that does not lose much accuracy while achieving very close to equality of odds (Hardt, et al., 2016). The method is flexible and applicable to multiple definitions of fairness as well as a wide range of gradient-based learning models, including both regression and classification tasks.
Automatic detection of toxic language plays an essential role in protecting social media users, especially minority groups, from verbal abuse. However, biases toward some attributes, including gender, race, and dialect, exist in most training datasets for toxicity detection. The biases make the learned models unfair and can even exacerbate the marginalization of people. Considering that current debiasing methods for general natural language understanding tasks cannot effectively mitigate the biases in the toxicity detectors, we propose to use invariant rationalization (InvRat), a game-theoretic framework consisting of a rationale generator and a predictor, to rule out the spurious correlation of certain syntactic patterns (e.g., identity mentions, dialect) to toxicity labels. We empirically show that our method yields lower false positive rate in both lexical and dialectal attributes than previous debiasing methods.
In this paper we derive a full expression for the propagation of weak lensing shape measurement biases into cosmic shear power spectra including the effect of missing data. We show using simulations that terms higher than first order in bias parameters can be ignored and the impact of biases can be captured by terms dependent only on the mean of the multiplicative bias field. We identify that the B-mode power contains information on the multiplicative bias. We find that without priors on the residual multiplicative bias $delta m$ and stochastic ellipticity variance $sigma_e$ that constraints on the amplitude of the cosmic shear power spectrum are completely degenerate, and that when applying priors the constrained amplitude $A$ is slightly biased low via a classic marginalisation paradox. Using all-sky Gaussian random field simulations we find that the combination of $(1+2delta m)A$ is unbiased for a joint EE and BB power spectrum likelihood if the error and mean (precision and accuracy) of the stochastic ellipticity variance is known to better than $sigma(sigma_e)leq 0.05$ and $Deltasigma_eleq 0.01$, or the multiplicative bias is known to better than $sigma(m)leq 0.07$ and $Delta mleq 0.01$.
As the statistical power of galaxy weak lensing reaches percent level precision, large, realistic and robust simulations are required to calibrate observational systematics, especially given the increased importance of object blending as survey depths increase. To capture the coupled effects of blending in both shear and photometric redshift calibration, we define the effective redshift distribution for lensing, $n_{gamma}(z)$, and describe how to estimate it using image simulations. We use an extensive suite of tailored image simulations to characterize the performance of the shear estimation pipeline applied to the Dark Energy Survey (DES) Year 3 dataset. We describe the multi-band, multi-epoch simulations, and demonstrate their high level of realism through comparisons to the real DES data. We isolate the effects that generate shear calibration biases by running variations on our fiducial simulation, and find that blending-related effects are the dominant contribution to the mean multiplicative bias of approximately $-2%$. By generating simulations with input shear signals that vary with redshift, we calibrate biases in our estimation of the effective redshfit distribution, and demonstrate the importance of this approach when blending is present. We provide corrected effective redshift distributions that incorporate statistical and systematic uncertainties, ready for use in DES Year 3 weak lensing analyses.