skip to main content
US FlagAn official website of the United States government
dot gov icon
Official websites use .gov
A .gov website belongs to an official government organization in the United States.
https lock icon
Secure .gov websites use HTTPS
A lock ( lock ) or https:// means you've safely connected to the .gov website. Share sensitive information only on official, secure websites.


Search for: All records

Award ID contains: 2328485

Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher. Some full text articles may not yet be available without a charge during the embargo (administrative interval).
What is a DOI Number?

Some links on this page may take you to non-federal websites. Their policies may differ from this site.

  1. ABSTRACT Stress drop is a fundamental parameter related to earthquake source physics, but is hard to measure accurately. To better understand how different factors influence stress-drop measurements, we compare two different methods using the Ridgecrest stress-drop validation data set: spectral decomposition (SD) and spectral ratio (SR), each with different processing options. We also examine the influence of spectral complexity on source parameter measurement. Applying the SD method, we find that frequency bandwidth and time-window length could influence spectral magnitude calibration, while depth-dependent attenuation is important to correctly map stress-drop variations. For the SR method, we find that the selected source model has limited influence on the measurements; however, the Boatwright model tends to produce smaller standard deviation and larger magnitude dependence than the Brune model. Variance reduction threshold, frequency bandwidth, and time-window length, if chosen within an appropriate parameter range, have limited influence on source parameter measurement. For both methods, wave type, attenuation correction, and spectral complexity strongly influence the result. The scale factor that quantifies the magnitude dependence of stress drop show large variations with different processing options, and earthquakes with complex source spectra deviating from the Brune-type source models tend to have larger scale factor than earthquakes without complexity. Based on these detailed comparisons, we make a few specific suggestions for data processing workflows that could help future studies of source parameters and interpretations. 
    more » « less
    Free, publicly-accessible full text available April 21, 2026
  2. ABSTRACT The recorded seismic waveform is a convolution of event source term, path term, and station term. Removing high-frequency attenuation due to path effect is a challenging problem. Empirical Green’s function (EGF) method uses nearly collocated small earthquakes to correct the path and station terms for larger events recorded at the same station. However, this method is subject to variability due to many factors. We focus on three events that were well recorded by the seismic network and a rapid response distributed acoustic sensing (DAS) array. Using a suite of high-quality EGF events, we assess the influence of time window, spectral measurement options, and types of data on the spectral ratio and relative source time function (RSTF) results. Increased number of tapers (from 2 to 16) tends to increase the measured corner frequency and reduce the source complexity. Extended long time window (e.g., 30 s) tends to produce larger variability of corner frequency. The multitaper algorithm that simultaneously optimizes both target and EGF spectra produces the most stable corner-frequency measurements. The stacked spectral ratio and RSTF from the DAS array are more stable than two nearby seismic stations, and are comparable to stacked results from the seismic network, suggesting that DAS array has strong potential in source characterization. 
    more » « less
    Free, publicly-accessible full text available March 4, 2026
  3. SUMMARY It is well known that large earthquakes often exhibit significant rupture complexity such as well separated subevents. With improved recording and data processing techniques, small earthquakes have been found to exhibit rupture complexity as well. Studying these small earthquakes offers the opportunity to better understand the possible causes of rupture complexities. Specifically, if they are random or are related to fault properties. We examine microearthquakes (M < 3) in the Parkfield, California, area that are recorded by a high-resolution borehole network. We quantify earthquake complexity by the deviation of source time functions and source spectra from simple circular (omega-square) source models. We establish thresholds to declare complexity, and find that it can be detected in earthquakes larger than magnitude 2, with the best resolution above M2.5. Comparison between the two approaches reveals good agreement (>90 per cent), implying both methods are characterizing the same source complexity. For the two methods, 60–80 per cent (M 2.6–3) of the resolved events are complex depending on the method. The complex events we observe tend to cluster in areas of previously identified structural complexity; a larger fraction of the earthquakes exhibit complexity in the days following the Mw 6 2004 Parkfield earthquake. Ignoring the complexity of these small events can introduce artefacts or add uncertainty to stress drop measurements. Focusing only on simple events however could lead to systematic bias, scaling artefacts and the lack of measurements of stress in structurally complex regions. 
    more » « less