No Arabic abstract
Despite the known general properties of the solar cycles, a reliable forecast of the 11-year sunspot number variations is still a problem. The difficulties are caused by the apparent chaotic behavior of the sunspot numbers from cycle to cycle and by the influence of various turbulent dynamo processes, which are far from understanding. For predicting the solar cycle properties we make an initial attempt to use the Ensemble Kalman Filter (EnKF), a data assimilation method, which takes into account uncertainties of a dynamo model and measurements, and allows to estimate future observational data. We present the results of forecasting of the solar cycles obtained by the EnKF method in application to a low-mode nonlinear dynamical system modeling the solar $alphaOmega$-dynamo process with variable magnetic helicity. Calculations of the predictions for the previous sunspot cycles show a reasonable agreement with the actual data. This forecast model predicts that the next sunspot cycle will be significantly weaker (by $sim 30%$) than the previous cycle, continuing the trend of low solar activity.
The prediction of solar flares, eruptions, and high energy particle storms is of great societal importance. The data mining approach to forecasting has been shown to be very promising. Benchmark datasets are a key element in the further development of data-driven forecasting. With one or more benchmark data sets established, judicious use of both the data themselves and the selection of prediction algorithms is key to developing a high quality and robust method for the prediction of geo-effective solar activity. We review here briefly the process of generating benchmark datasets and developing prediction algorithms.
The use of data assimilation technique to identify optimal topography is discussed in frames of time-dependent motion governed by non-linear barotropic ocean model. Assimilation of artificially generated data allows to measure the influence of various error sources and to classify the impact of noise that is present in observational data and model parameters. The choice of assimilation window is discussed. Assimilating noisy data with longer windows provides higher accuracy of identified topography. The topography identified once by data assimilation can be successfully used for other model runs that start from other initial conditions and are situated in other parts of the models attractor.
We develop a physics-informed machine learning approach for large-scale data assimilation and parameter estimation and apply it for estimating transmissivity and hydraulic head in the two-dimensional steady-state subsurface flow model of the Hanford Site given synthetic measurements of said variables. In our approach, we extend the physics-informed conditional Karhunen-Lo{e}ve expansion (PICKLE) method for modeling subsurface flow with unknown flux (Neumann) and varying head (Dirichlet) boundary conditions. We demonstrate that the PICKLE method is comparable in accuracy with the standard maximum a posteriori (MAP) method, but is significantly faster than MAP for large-scale problems. Both methods use a mesh to discretize the computational domain. In MAP, the parameters and states are discretized on the mesh; therefore, the size of the MAP parameter estimation problem directly depends on the mesh size. In PICKLE, the mesh is used to evaluate the residuals of the governing equation, while the parameters and states are approximated by the truncated conditional Karhunen-Lo{e}ve expansions with the number of parameters controlled by the smoothness of the parameter and state fields, and not by the mesh size. For a considered example, we demonstrate that the computational cost of PICKLE increases near linearly (as $N_{FV}^{1.15}$) with the number of grid points $N_{FV}$, while that of MAP increases much faster as $N_{FV}^{3.28}$. We demonstrated that once trained for one set of Dirichlet boundary conditions (i.e., one river stage), the PICKLE method provides accurate estimates of the hydraulic head for any value of the Dirichlet boundary conditions (i.e., for any river stage).
The Sun exhibits a well-observed modulation in the number of spots on its disk over a period of about 11 years. From the dawn of modern observational astronomy sunspots have presented a challenge to understanding -- their quasi-periodic variation in number, first noted 175 years ago, stimulates community-wide interest to this day. A large number of techniques are able to explain the temporal landmarks, (geometric) shape, and amplitude of sunspot cycles, however forecasting these features accurately in advance remains elusive. Recent observationally-motivated studies have illustrated a relationship between the Suns 22-year (Hale) magnetic cycle and the production of the sunspot cycle landmarks and patterns, but not the amplitude of the sunspot cycle. Using (discrete) Hilbert transforms on more than 270 years of (monthly) sunspot numbers we robustly identify the so-called termination events that mark the end of the previous 11-yr sunspot cycle, the enhancement/acceleration of the present cycle, and the end of 22-yr magnetic activity cycles. Using these we extract a relationship between the temporal spacing of terminators and the magnitude of sunspot cycles. Given this relationship and our prediction of a terminator event in 2020, we deduce that Sunspot Cycle 25 could have a magnitude that rivals the top few since records began. This outcome would be in stark contrast to the community consensus estimate of sunspot cycle 25 magnitude.
Data assimilation leads naturally to a Bayesian formulation in which the posterior probability distribution of the system state, given the observations, plays a central conceptual role. The aim of this paper is to use this Bayesian posterior probability distribution as a gold standard against which to evaluate various commonly used data assimilation algorithms. A key aspect of geophysical data assimilation is the high dimensionality and low predictability of the computational model. With this in mind, yet with the goal of allowing an explicit and accurate computation of the posterior distribution, we study the 2D Navier-Stokes equations in a periodic geometry. We compute the posterior probability distribution by state-of-the-art statistical sampling techniques. The commonly used algorithms that we evaluate against this accurate gold standard, as quantified by comparing the relative error in reproducing its moments, are 4DVAR and a variety of sequential filtering approximations based on 3DVAR and on extended and ensemble Kalman filters. The primary conclusions are that: (i) with appropriate parameter choices, approximate filters can perform well in reproducing the mean of the desired probability distribution; (ii) however they typically perform poorly when attempting to reproduce the covariance; (iii) this poor performance is compounded by the need to modify the covariance, in order to induce stability. Thus, whilst filters can be a useful tool in predicting mean behavior, they should be viewed with caution as predictors of uncertainty. These conclusions are intrinsic to the algorithms and will not change if the model complexity is increased, for example by employing a smaller viscosity, or by using a detailed NWP model.