No Arabic abstract
We present a novel algorithm that predicts the probability that the time derivative of the horizontal component of the ground magnetic field $dB/dt$ exceeds a specified threshold at a given location. This quantity provides important information that is physically relevant to Geomagnetically Induced Currents (GIC), which are electric currents { associated to} sudden changes in the Earths magnetic field due to Space Weather events. The model follows a gray-box approach by combining the output of a physics-based model with machine learning. Specifically, we combine the University of Michigans Geospace model that is operational at the NOAA Space Weather Prediction Center, with a boosted ensemble of classification trees. We discuss the problem of re-calibrating the output of the decision tree to obtain reliable probabilities. The performance of the model is assessed by typical metrics for probabilistic forecasts: Probability of Detection and False Detection, True Skill Statistic, Heidke Skill Score, and Receiver Operating Characteristic curve. We show that the ML enhanced algorithm consistently improves all the metrics considered.
Motivated by low-altitude cusp observations of small-scale (~ 1 km) field-aligned currents (SSFACs) interpreted as ionospheric Alfven resonator modes, we investigated the effects of Alfven wave energy deposition on thermospheric upwelling and the formation of air density enhancements in and near the cusp. Such density enhancements were commonly observed near 400 km altitude by the CHAMP satellite. They are not predicted by empirical thermosphere models, and they are well-correlated with the observed SSFACs. A parameterized model for the altitude dependence of the Alfven wave electric field, constrained by CHAMP data, has been developed and embedded in the Joule heating module of the National Center for Atmospheric Research (NCAR) Coupled Magnetosphere-Ionosphere-Thermosphere (CMIT) model. The CMIT model was then used to simulate the geospace response to an interplanetary stream interaction region (SIR) that swept past Earth on 26-27 March 2003. CMIT diagnostics for the thermospheric mass density at 400 km altitude show: 1) CMIT without Alfvenic Joule heating usually underestimates CHAMPs orbit-average density; inclusion of Alfvenic heating modestly improves CMITs orbit-average prediction of the density (by a few %), especially during the more active periods of the SIR event. 2) The improvement in CMITs instantaneous density prediction with Alfvenic heating included is more significant (up to 15%) in the vicinity of the cusp heating region, a feature that the MSIS empirical thermosphere model misses for this event. Thermospheric density changes of 20-30% caused by the cusp-region Alfvenic heating sporadically populate the polar region through the action of corotation and neutral winds.
We advance the modeling capability of electron particle precipitation from the magnetosphere to the ionosphere through a new database and use of machine learning (ML) tools to gain utility from those data. We have compiled, curated, analyzed, and made available a new and more capable database of particle precipitation data that includes 51 satellite years of Defense Meteorological Satellite Program (DMSP) observations temporally aligned with solar wind and geomagnetic activity data. The new total electron energy flux particle precipitation nowcast model, a neural network called PrecipNet, takes advantage of increased expressive power afforded by ML approaches to appropriately utilize diverse information from the solar wind and geomagnetic activity and, importantly, their time histories. With a more capable representation of the organizing parameters and the target electron energy flux observations, PrecipNet achieves a >50% reduction in errors from a current state-of-the-art model oval variation, assessment, tracking, intensity, and online nowcasting (OVATION Prime), better captures the dynamic changes of the auroral flux, and provides evidence that it can capably reconstruct mesoscale phenomena. We create and apply a new framework for space weather model evaluation that culminates previous guidance from across the solar-terrestrial research community. The research approach and results are representative of the new frontier of space weather research at the intersection of traditional and data science-driven discovery and provides a foundation for future efforts.
The forecast of the time of arrival of a coronal mass ejection (CME) to Earth is of critical importance for our high-technology society and for any future manned exploration of the Solar System. As critical as the forecast accuracy is the knowledge of its precision, i.e. the error associated to the estimate. We propose a statistical approach for the computation of the time of arrival using the drag-based model by introducing the probability distributions, rather than exact values, as input parameters, thus allowing the evaluation of the uncertainty on the forecast. We test this approach using a set of CMEs whose transit times are known, and obtain extremely promising results: the average value of the absolute differences between measure and forecast is 9.1h, and half of these residuals are within the estimated errors. These results suggest that this approach deserves further investigation. We are working to realize a real-time implementation which ingests the outputs of automated CME tracking algorithms as inputs to create a database of events useful for a further validation of the approach.
In Bayesian classification, it is important to establish a probabilistic model for each class for likelihood estimation. Most of the previous methods modeled the probability distribution in the whole sample space. However, real-world problems are usually too complex to model in the whole sample space; some fundamental assumptions are required to simplify the global model, for example, the class conditional independence assumption for naive Bayesian classification. In this paper, with the insight that the distribution in a local sample space should be simpler than that in the whole sample space, a local probabilistic model established for a local region is expected much simpler and can relax the fundamental assumptions that may not be true in the whole sample space. Based on these advantages we propose establishing local probabilistic models for Bayesian classification. In addition, a Bayesian classifier adopting a local probabilistic model can even be viewed as a generalized local classification model; by tuning the size of the local region and the corresponding local model assumption, a fitting model can be established for a particular classification problem. The experimental results on several real-world datasets demonstrate the effectiveness of local probabilistic models for Bayesian classification.
We perform a probabilistic analysis of onion routing. The analysis is presented in a black-box model of anonymous communication in the Universally Composable framework that abstracts the essential properties of onion routing in the presence of an active adversary that controls a portion of the network and knows all a priori distributions on user choices of destination. Our results quantify how much the adversary can gain in identifying users by exploiting knowledge of their probabilistic behavior. In particular, we show that, in the limit as the network gets large, a user us anonymity is worst either when the other users always choose the destination u is least likely to visit or when the other users always choose the destination u chooses. This worst-case anonymity with an adversary that controls a fraction b of the routers is shown to be comparable to the best-case anonymity against an adversary that controls a fraction surdb.