Do you want to publish a course? Click here

Static behaviour of induced seismicity

60   0   0.0 ( 0 )
 Added by Arnaud Mignan
 Publication date 2015
  fields Physics
and research's language is English
 Authors Arnaud Mignan




Ask ChatGPT about the research

The standard paradigm to describe seismicity induced by fluid injection is to apply nonlinear diffusion dynamics in a poroelastic medium. I show that the spatiotemporal behaviour and rate evolution of induced seismicity can, instead, be expressed by geometric operations on a static stress field produced by volume change at depth. I obtain laws similar in form to the ones derived from poroelasticity while requiring a lower description length. Although fluid flow is known to occur in the ground, it is not pertinent to the behaviour of induced seismicity. The proposed model is equivalent to the static stress model for tectonic foreshocks generated by the Non- Critical Precursory Accelerating Seismicity Theory. This study hence verifies the explanatory power of this theory outside of its original scope.



rate research

Read More

The rise in the frequency of anthropogenic earthquakes due to deep fluid injections is posing serious economic, societal, and legal challenges to geo-energy and waste-disposal projects. We propose an actuarial approach to mitigate this risk, first by defining an autonomous decision-making process based on an adaptive traffic light system (ATLS) to stop risky injections, and second by quantifying a cost of public safety based on the probability of an injection-well being abandoned. The ATLS underlying statistical model is first confirmed to be representative of injection-induced seismicity, with examples taken from past reservoir stimulation experiments (mostly from Enhanced Geothermal Systems, EGS). Then the decision strategy is formalized: Being integrable, the model yields a closed-form ATLS solution that maps a risk-based safety standard or norm to an earthquake magnitude not to exceed during stimulation. Finally, the EGS levelized cost of electricity (LCOE) is reformulated in terms of null expectation, with the cost of abandoned injection-well implemented. We find that the price increase to mitigate the increased seismic risk in populated areas can counterbalance the heat credit. However this public safety cost disappears if buildings are based on earthquake-resistant designs or if a more relaxed risk safety standard or norm is chosen.
Several recent works point out that the crowd of small unobservable earthquakes (with magnitudes below the detection threshold $m_d$) may play a significant and perhaps dominant role in triggering future seismicity. Using the ETAS branching model of triggered seismicity, we apply the formalism of generating probability functions to investigate how the statistical properties of observable earthquakes differ from the statistics of all events. The ETAS (epidemic-type aftershock sequence) model assumes that each earthquake can trigger other earthquakes (``aftershocks). An aftershock sequence results in this model from the cascade of aftershocks of each past earthquake. The triggering efficiency of earthquakes is assumed to vanish below a lower magnitude limit $m_0$, in order to ensure the convergence of the theory and may reflect the physics of state-and-velocity frictional rupture. We show that, to a good approximation, the ETAS model is renormalized onto itself under what amounts to a decimation procedure $m_0 to m_d$, with just a renormalization of the branching ratio from $n$ to an effective value $n(m_d)$. Our present analysis thus confirms, for the full statistical properties, the results obtained previously by one of us and Werner, based solely on the average seismic rates (the first-order moment of the statistics). However, our analysis also demonstrates that this renormalization is not exact, as there are small corrections which can be systematically calculated, in terms of additional contributions that can be mapped onto a different branching model (a new relevant direction in the language of the renormalization group).
176 - D. Sornette 2008
This entry in the Encyclopedia of Complexity and Systems Science, Springer present a summary of some of the concepts and calculational tools that have been developed in attempts to apply statistical physics approaches to seismology. We summarize the leading theoretical physical models of the space-time organization of earthquakes. We present a general discussion and several examples of the new metrics proposed by statistical physicists, underlining their strengths and weaknesses. The entry concludes by briefly outlining future directions. The presentation is organized as follows. I Glossary II Definition and Importance of the Subject III Introduction IV Concepts and Calculational Tools IV.1 Renormalization, Scaling and the Role of Small Earthquakes in Models of Triggered Seismicity IV.2 Universality IV.3 Intermittent Periodicity and Chaos IV.4 Turbulence IV.5 Self-Organized Criticality V Competing mechanisms and models V.1 Roots of complexity in seismicity: dynamics or heterogeneity? V.2 Critical earthquakes V.3 Spinodal decomposition V.4 Dynamics, stress interaction and thermal fluctuation effects VI Empirical studies of seismicity inspired by statistical physics VI.1 Early successes and latter subsequent challenges VI.2 Entropy method for the distribution of time intervals between mainshocks VI.3 Scaling of the PDF of Waiting Times VI.4 Scaling of the PDF of Distances Between Subsequent Earthquakes VI.5 The Network Approach VII Future Directions
Seismicity and faulting within the Earth crust are characterized by many scaling laws that are usually interpreted as qualifying the existence of underlying physical mechanisms associated with some kind of criticality in the sense of phase transitions. Using an augmented Epidemic-Type Aftershock Sequence (ETAS) model that accounts for the spatial variability of the background rates $mu(x,y)$, we present a direct quantitative test of criticality. We calibrate the model to the ANSS catalog of the entire globe, the region around California, and the Geonet catalog for the region around New Zealand using an extended Expectation-Maximization (EM) algorithm including the determination of $mu(x,y)$. We demonstrate that the criticality reported in previous studies is spurious and can be attributed to a systematic upward bias in the calibration of the branching ratio of the ETAS model, when not accounting correctly for spatial variability. We validate the version of the ETAS model which possesses a space varying background rate $mu(x,y)$ by performing pseudo prospective forecasting tests. The non-criticality of seismicity has major implications for the prediction of large events.
We report a similarity of fluctuations in equilibrium critical phenomena and non-equilibrium systems, which is based on the concept of natural time. The world-wide seismicity as well as that of San Andreas fault system and Japan are analyzed. An order parameter is chosen and its fluctuations relative to the standard deviation of the distribution are studied. We find that the scaled distributions fall on the same curve, which interestingly exhibits, over four orders of magnitude, features similar to those in several equilibrium critical phenomena (e.g., 2D Ising model) as well as in non-equilibrium systems (e.g., 3D turbulent flow).
comments
Fetching comments Fetching comments
Sign in to be able to follow your search criteria
mircosoft-partner

هل ترغب بارسال اشعارات عن اخر التحديثات في شمرا-اكاديميا