ترغب بنشر مسار تعليمي؟ اضغط هنا

Statistical Physics Approaches to Seismicity

124   0   0.0 ( 0 )
 نشر من قبل Didier Sornette
 تاريخ النشر 2008
  مجال البحث فيزياء
والبحث باللغة English
 تأليف D. Sornette




اسأل ChatGPT حول البحث

This entry in the Encyclopedia of Complexity and Systems Science, Springer present a summary of some of the concepts and calculational tools that have been developed in attempts to apply statistical physics approaches to seismology. We summarize the leading theoretical physical models of the space-time organization of earthquakes. We present a general discussion and several examples of the new metrics proposed by statistical physicists, underlining their strengths and weaknesses. The entry concludes by briefly outlining future directions. The presentation is organized as follows. I Glossary II Definition and Importance of the Subject III Introduction IV Concepts and Calculational Tools IV.1 Renormalization, Scaling and the Role of Small Earthquakes in Models of Triggered Seismicity IV.2 Universality IV.3 Intermittent Periodicity and Chaos IV.4 Turbulence IV.5 Self-Organized Criticality V Competing mechanisms and models V.1 Roots of complexity in seismicity: dynamics or heterogeneity? V.2 Critical earthquakes V.3 Spinodal decomposition V.4 Dynamics, stress interaction and thermal fluctuation effects VI Empirical studies of seismicity inspired by statistical physics VI.1 Early successes and latter subsequent challenges VI.2 Entropy method for the distribution of time intervals between mainshocks VI.3 Scaling of the PDF of Waiting Times VI.4 Scaling of the PDF of Distances Between Subsequent Earthquakes VI.5 The Network Approach VII Future Directions

قيم البحث

اقرأ أيضاً

We show how two techniques from statistical physics can be adapted to solve a variant of the notorious Unique Games problem, potentially opening new avenues towards the Unique Games Conjecture. The variant, which we call Count Unique Games, is a prom ise problem in which the yes case guarantees a certain number of highly satisfiable assignments to the Unique Games instance. In the standard Unique Games problem, the yes case only guarantees at least one such assignment. We exhibit efficient algorithms for Count Unique Games based on approximating a suitable partition function for the Unique Games instance via (i) a zero-free region and polynomial interpolation, and (ii) the cluster expansion. We also show that a modest improvement to the parameters for which we give results would refute the Unique Games Conjecture.
Several recent works point out that the crowd of small unobservable earthquakes (with magnitudes below the detection threshold $m_d$) may play a significant and perhaps dominant role in triggering future seismicity. Using the ETAS branching model of triggered seismicity, we apply the formalism of generating probability functions to investigate how the statistical properties of observable earthquakes differ from the statistics of all events. The ETAS (epidemic-type aftershock sequence) model assumes that each earthquake can trigger other earthquakes (``aftershocks). An aftershock sequence results in this model from the cascade of aftershocks of each past earthquake. The triggering efficiency of earthquakes is assumed to vanish below a lower magnitude limit $m_0$, in order to ensure the convergence of the theory and may reflect the physics of state-and-velocity frictional rupture. We show that, to a good approximation, the ETAS model is renormalized onto itself under what amounts to a decimation procedure $m_0 to m_d$, with just a renormalization of the branching ratio from $n$ to an effective value $n(m_d)$. Our present analysis thus confirms, for the full statistical properties, the results obtained previously by one of us and Werner, based solely on the average seismic rates (the first-order moment of the statistics). However, our analysis also demonstrates that this renormalization is not exact, as there are small corrections which can be systematically calculated, in terms of additional contributions that can be mapped onto a different branching model (a new relevant direction in the language of the renormalization group).
59 - Arnaud Mignan 2015
The standard paradigm to describe seismicity induced by fluid injection is to apply nonlinear diffusion dynamics in a poroelastic medium. I show that the spatiotemporal behaviour and rate evolution of induced seismicity can, instead, be expressed by geometric operations on a static stress field produced by volume change at depth. I obtain laws similar in form to the ones derived from poroelasticity while requiring a lower description length. Although fluid flow is known to occur in the ground, it is not pertinent to the behaviour of induced seismicity. The proposed model is equivalent to the static stress model for tectonic foreshocks generated by the Non- Critical Precursory Accelerating Seismicity Theory. This study hence verifies the explanatory power of this theory outside of its original scope.
Seismicity and faulting within the Earth crust are characterized by many scaling laws that are usually interpreted as qualifying the existence of underlying physical mechanisms associated with some kind of criticality in the sense of phase transition s. Using an augmented Epidemic-Type Aftershock Sequence (ETAS) model that accounts for the spatial variability of the background rates $mu(x,y)$, we present a direct quantitative test of criticality. We calibrate the model to the ANSS catalog of the entire globe, the region around California, and the Geonet catalog for the region around New Zealand using an extended Expectation-Maximization (EM) algorithm including the determination of $mu(x,y)$. We demonstrate that the criticality reported in previous studies is spurious and can be attributed to a systematic upward bias in the calibration of the branching ratio of the ETAS model, when not accounting correctly for spatial variability. We validate the version of the ETAS model which possesses a space varying background rate $mu(x,y)$ by performing pseudo prospective forecasting tests. The non-criticality of seismicity has major implications for the prediction of large events.
The rise in the frequency of anthropogenic earthquakes due to deep fluid injections is posing serious economic, societal, and legal challenges to geo-energy and waste-disposal projects. We propose an actuarial approach to mitigate this risk, first by defining an autonomous decision-making process based on an adaptive traffic light system (ATLS) to stop risky injections, and second by quantifying a cost of public safety based on the probability of an injection-well being abandoned. The ATLS underlying statistical model is first confirmed to be representative of injection-induced seismicity, with examples taken from past reservoir stimulation experiments (mostly from Enhanced Geothermal Systems, EGS). Then the decision strategy is formalized: Being integrable, the model yields a closed-form ATLS solution that maps a risk-based safety standard or norm to an earthquake magnitude not to exceed during stimulation. Finally, the EGS levelized cost of electricity (LCOE) is reformulated in terms of null expectation, with the cost of abandoned injection-well implemented. We find that the price increase to mitigate the increased seismic risk in populated areas can counterbalance the heat credit. However this public safety cost disappears if buildings are based on earthquake-resistant designs or if a more relaxed risk safety standard or norm is chosen.
التعليقات
جاري جلب التعليقات جاري جلب التعليقات
سجل دخول لتتمكن من متابعة معايير البحث التي قمت باختيارها
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا