Do you want to publish a course? Click here

Embracing Data Incompleteness for Better Earthquake Forecasting

96   0   0.0 ( 0 )
 Added by Leila Mizrahi
 Publication date 2021
  fields Physics
and research's language is English




Ask ChatGPT about the research

We propose two new methods to calibrate the parameters of the Epidemic-Type Aftershock Sequence (ETAS) model based on expectation maximization (EM) while accounting for temporal variation of catalog completeness. The first method allows for model calibration on earthquake catalogs with long history, featuring temporal variation of the magnitude of completeness, $m_c$. This extended calibration technique is beneficial for long-term Probabilistic Seismic Hazard Assessment (PSHA), which is often based on a mixture of instrumental and historical catalogs. The second method jointly estimates ETAS parameters and high-frequency detection incompleteness to address the potential biases in parameter calibration due to short-term aftershock incompleteness. For this, we generalize the concept of completeness magnitude and consider a rate- and magnitude-dependent detection probability $-$ embracing incompleteness instead of avoiding it. Using synthetic tests, we show that both methods can accurately invert the parameters of simulated catalogs. We then use them to estimate ETAS parameters for California using the earthquake catalog since 1932. To explore how the newly gained information from the second method affects earthquakes predictability, we conduct pseudo-prospective forecasting experiments for California. Our proposed model significantly outperforms the base ETAS model, and we find that the ability to include small earthquakes for simulation of future scenarios is the main driver of the improvement. Our results point towards a preference of earthquakes to trigger similarly sized aftershocks, which has potentially major implications for our understanding of earthquake interaction mechanisms and for the future of seismicity forecasting.



rate research

Read More

Forecasting the full distribution of the number of earthquakes is revealed to be inherently superior to forecasting their mean. Forecasting the full distribution of earthquake numbers is also shown to yield robust projections in the presence of surprise large earthquakes, which in the past have strongly deteriorated the scores of existing models. We show this with pseudo-prospective experiments on synthetic as well as real data from the Advanced National Seismic System (ANSS) database for California, with earthquakes with magnitude larger than 2.95 that occurred between the period 1971-2016. Our results call in question the testing methodology of the Collaboratory for the study of earthquake predictability (CSEP), which amounts to assuming a Poisson distribution of earthquake numbers, which is known to be a poor representation of the heavy-tailed distribution of earthquake numbers. Using a spatially varying ETAS model, we demonstrate a remarkable stability of the forecasting performance, when using the full distribution of earthquake numbers for the forecasts, even in the presence of large earthquakes such as Mw 7.1 Hector Mine, Mw 7.2 El Mayor-Cucapah, Mw 6.6 Sam Simeon earthquakes, or in the presence of intense swarm activity in Northwest Nevada in 2014. While our results have been derived for ETAS type models, we propose that all earthquake forecasting models of any type should embrace the full distribution of earthquake numbers, such that their true forecasting potential is revealed.
Currently, one of the best performing and most popular earthquake forecasting models rely on the working hypothesis that: locations of past background earthquakes reveal the probable location of future seismicity. As an alternative, we present a class of smoothed seismicity models (SSMs) based on the principles of the Epidemic Type Aftershock Sequence (ETAS) model, which forecast the location, time and magnitude of all future earthquakes using the estimates of the background seismicity rate and the rates of future aftershocks of all generations. Using the Californian earthquake catalog, we formulate six controlled pseudo-prospective experiments with different combination of three target magnitude thresholds: 2.95, 3.95 or 4.95 and two forecasting time horizons: 1 or 5 year. In these experiments, we compare the performance of:(1) ETAS model with spatially homogenous parameters or GETAS (2) ETAS model with spatially variable parameters or SVETAS (3) three declustering based SSMs (4) a simple SSM based on undeclustered data and (5) a model based on strain rate data, in forecasting the location and magnitude of all (undeclustered) target earthquakes during many testing periods. In all conducted experiments, the SVETAS model comes out with consistent superiority compared to all the competing models. Consistently better performance of SVETAS model with respect to declustering based SSMs highlights the importance of forecasting the future aftershocks of all generations for developing better earthquake forecasting models. Among the two ETAS models themselves, accounting for the optimal spatial variation of the parameters leads to strong and statistically significant improvements in forecasting performance.
We examine the precursory behavior of geoelectric signals before large earthquakes by means of an algorithm including an alarm-based model and binary classification. This algorithm, introduced originally by Chen and Chen [Nat. Hazards., 84, 2016], is improved by removing a time parameter for coarse-graining of earthquake occurrences, as well as by extending the single station method into a joint stations method. We also determine the optimal frequency bands of earthquake-related geoelectric signals with the highest signal-to-noise ratio. Using significance tests, we also provide evidence of an underlying seismoelectric relationship. It is appropriate for machine learning to extract this underlying relationship, which could be used to quantify probabilistic forecasts of impending earthquakes, and to get closer to operational earthquake prediction.
133 - Sumiyoshi Abe 2010
Earthquake network is known to be of the small-world type. The values of the network characteristics, however, depend not only on the cell size (i.e., the scale of coarse graining needed for constructing the network) but also on the size of a seismic data set. Here, discovery of a scaling law for the clustering coefficient in terms of the data size, which is refereed to here as finite data-size scaling, is reported. Its universality is shown to be supported by the detailed analysis of the data taken from California, Japan and Iran. Effects of setting threshold of magnitude are also discussed.
An article for the Springer Encyclopedia of Complexity and System Science
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا