No Arabic abstract
Robust machine learning relies on access to data that can be used with standardized frameworks in important tasks and the ability to develop models whose performance can be reasonably reproduced. In machine learning for healthcare, the community faces reproducibility challenges due to a lack of publicly accessible data and a lack of standardized data processing frameworks. We present MIMIC-Extract, an open-source pipeline for transforming raw electronic health record (EHR) data for critical care patients contained in the publicly-available MIMIC-III database into dataframes that are directly usable in common machine learning pipelines. MIMIC-Extract addresses three primary challenges in making complex health records data accessible to the broader machine learning community. First, it provides standardized data processing functions, including unit conversion, outlier detection, and aggregating semantically equivalent features, thus accounting for duplication and reducing missingness. Second, it preserves the time series nature of clinical data and can be easily integrated into clinically actionable prediction tasks in machine learning for health. Finally, it is highly extensible so that other researchers with related questions can easily use the same pipeline. We demonstrate the utility of this pipeline by showcasing several benchmark tasks and baseline results.
Electronic health records (EHR) systems contain vast amounts of medical information about patients. These data can be used to train machine learning models that can predict health status, as well as to help prevent future diseases or disabilities. However, getting patients medical data to obtain well-trained machine learning models is a challenging task. This is because sharing the patients medical records is prohibited by law in most countries due to patients privacy concerns. In this paper, we tackle this problem by sharing the models instead of the original sensitive data by using the mimic learning approach. The idea is first to train a model on the original sensitive data, called the teacher model. Then, using this model, we can transfer its knowledge to another model, called the student model, without the need to learn the original data used in training the teacher model. The student model is then shared to the public and can be used to make accurate predictions. To assess the mimic learning approach, we have evaluated our scheme using different medical datasets. The results indicate that the student model mimics the teacher model performance in terms of prediction accuracy without the need to access to the patients original data records.
Multiple Indicators Multiple Causes (MIMIC) models are type of structural equation models, a theory-based approach to confirm the influence of a set of exogenous causal variables on the latent variable, and also the effect of the latent variable on observed indicator variables. In a common MIMIC model, multiple indicators reflect the underlying latent variables/factors, and the multiple causes (observed predictors) affect latent variables/factors. Basic assumptions of MIMIC are clearly violated in case of a variable being both an indicator and a cause, i.e. in the presence of reverse causality. Furthermore, the model is then unidentified. To resolve the situation, which can arise frequently, and as MIMIC estimation lacks closed form solutions for parameters we utilize a version of Bollens (1996) 2SLS estimator for structural equation models combined with Joreskog (1970)s method of the analysis of covariance structures to derive a new, 2SLS estimator for MIMIC models. Our 2SLS empirical estimation is based on static MIMIC specification but we point also to dynamic/error-correction MIMIC specification and 2SLS solution for it. We derive basic asymptotic theory for static 2SLS-MIMIC, present a simulation study and apply findings to an interesting empirical case of estimating precarious status of older workers (using dataset of Survey of Health, Ageing and Retirement in Europe) which solves an important issue of the definition of precarious work as a multidimensional concept, not modelled adequately so far.
The ability to act in multiple environments and transfer previous knowledge to new situations can be considered a critical aspect of any intelligent agent. Towards this goal, we define a novel method of multitask and transfer learning that enables an autonomous agent to learn how to behave in multiple tasks simultaneously, and then generalize its knowledge to new domains. This method, termed Actor-Mimic, exploits the use of deep reinforcement learning and model compression techniques to train a single policy network that learns how to act in a set of distinct tasks by using the guidance of several expert teachers. We then show that the representations learnt by the deep policy network are capable of generalizing to new tasks with no prior expert guidance, speeding up learning in novel environments. Although our method can in general be applied to a wide range of problems, we use Atari games as a testing environment to demonstrate these methods.
Interest in magnetic-tunnel junctions has prompted a re-examination of tunneling measurements through thin insulating films. In any study of metal-insulator-metal trilayers, one tries to eliminate the possibility of pinholes (small areas over which the thickness of the insulator goes to zero so that the upper and lower metals of the trilayer make direct contact). Recently, we have presented experimental evidence that ferromagnet-insulator-normal trilayers that appear from current-voltage plots to be pinhole-free may nonetheless in some cases harbor pinholes. Here, we show how pinholes may arise in a simple but realistic model of film deposition and that purely classical conduction through pinholes may mimic one aspect of tunneling, the exponential decay in current with insulating thickness.
At present, adversarial attacks are designed in a task-specific fashion. However, for downstream computer vision tasks such as image captioning, image segmentation etc., the current deep learning systems use an image classifier like VGG16, ResNet50, Inception-v3 etc. as a feature extractor. Keeping this in mind, we propose Mimic and Fool, a task agnostic adversarial attack. Given a feature extractor, the proposed attack finds an adversarial image which can mimic the image feature of the original image. This ensures that the two images give the same (or similar) output regardless of the task. We randomly select 1000 MSCOCO validation images for experimentation. We perform experiments on two image captioning models, Show and Tell, Show Attend and Tell and one VQA model, namely, end-to-end neural module network (N2NMN). The proposed attack achieves success rate of 74.0%, 81.0% and 87.1% for Show and Tell, Show Attend and Tell and N2NMN respectively. We also propose a slight modification to our attack to generate natural-looking adversarial images. In addition, we also show the applicability of the proposed attack for invertible architecture. Since Mimic and Fool only requires information about the feature extractor of the model, it can be considered as a gray-box attack.