ArticlePDF Available

Improvements in Data Quality, Integration and Reliability: New Developments at the IRIS DMC

Authors:

Abstract and Figures

With the support of the US National Science Foundation (NSF) and on behalf of the international seismological community, IRIS developed a Data Management Center (DMC; Ahern, 2003) that has for decades acted as a primary resource for seismic networks wishing to make their data broadly available, as well as a significant point of access for researchers and monitoring agencies worldwide that wish to access high quality data for a variety of purposes. Recently IRIS has taken significant new steps to improve the quality of and access to the services of the IRIS DMC. This paper highlights some of the current new efforts being undertaken by IRIS. The primary topics include (1) steps to improve reliability and consistency of access to IRIS data resources, (2) a comprehensive new approach to assessing the quality of seismological and other data, (3) working with international partners to federate seismological data access services, and finally (4) extensions of the federated concept to extend data access to data from other geoscience domains.
Content may be subject to copyright.
Adv. Geosci., 40, 31–35, 2015
www.adv-geosci.net/40/31/2015/
doi:10.5194/adgeo-40-31-2015
© Author(s) 2015. CC Attribution 3.0 License.
Improvements in Data Quality, Integration and Reliability:
New Developments at the IRIS DMC
T. Ahern, R. Benson, R. Casey, C. Trabant, and B. Weertman
IRIS Data Management Center, 1408 NE 45th Street, Seattle, WA 98105, USA
Correspondence to: T. Ahern (tim@iris.washington.edu)
Received: 14 November 2014 – Revised: 14 January 2015 – Accepted: 15 January 2015 – Published: 13 February 2015
Abstract. With the support of the US National Science Foun-
dation (NSF) and on behalf of the international seismologi-
cal community, IRIS developed a Data Management Center
(DMC; Ahern, 2003) that has for decades acted as a pri-
mary resource for seismic networks wishing to make their
data broadly available, as well as a significant point of access
for researchers and monitoring agencies worldwide that wish
to access high quality data for a variety of purposes. Recently
IRIS has taken significant new steps to improve the quality
of and access to the services of the IRIS DMC. This paper
highlights some of the current new efforts being undertaken
by IRIS. The primary topics include (1) steps to improve re-
liability and consistency of access to IRIS data resources,
(2) a comprehensive new approach to assessing the quality of
seismological and other data, (3) working with international
partners to federate seismological data access services, and
finally (4) extensions of the federated concept to extend data
access to data from other geoscience domains.
1 Building resiliency in IRIS data services through
auxiliary data centers
For several decades IRIS has relied on a single centralized
data center in Seattle, Washington to provide all services to
the community. In 2006 we created an Active Backup Sys-
tem at the PASSCAL Instrument Center in Socorro, New
Mexico. This backup system held copies of all the primary
waveform data, key software source and binaries, documen-
tation and a variety of other information in case of a catas-
trophic event at the primary center such as fire or earthquake.
In 2009 the system was relocated to the UNAVCO facility
in Boulder, Colorado to take advantage of the higher band-
width at that location. In 2013 the DMC transitioned the Ac-
tive Backup concept to a fully functioning Auxiliary Data
Center (ADC) where ultimately all of the services of the pri-
mary IRIS data center would be replicated and available at
all times. The first auxiliary center is located at the Liver-
more Valley Open Campus (LVOC) at Lawrence Livermore
National Laboratory (LLNL). The location of the ADC al-
lows it to be connected to the High Performance Comput-
ing (HPC) environment at LLNL and helps in IRIS’ goal to
place the entire IRIS archive in the proximity of supercom-
puting resources. The key development that enabled IRIS to
do this relies on the Service Oriented Architecture (SOA) that
IRIS has developed. The replication of the DMC functional-
ity was greatly aided by relying on web services that have
been adopted by the International Federation of Digital Seis-
mograph Networks (FDSN) as well as a small number of ad-
ditional services that IRIS has developed and uses internally
as well as exposing them external to the DMC.
Most of IRIS’ applications use these web services and it
is much simpler to deploy systems at multiple locations once
the web service infrastructure is deployed. This infrastructure
is fully deployed at the ADC and the ADC services function
identically to the services at the primary DMC.
In the future it is IRIS’ plan to leverage an external load
balancing system that will seamlessly route some requests to
the primary DMC and others to the ADC based on business
rules such as how busy one system is over the other system,
geographic proximity to one or the other services, or other
business rules yet to be determined. Currently the global load
balancing is not in place. It is possible to access the web ser-
vices and other installed applications at the ADC if the URL
is known. In fact we are currently running the MUSTANG
QA system at both locations leveraging the web services in-
ternally so the proof of concept has been heavily exercised.
Specifically IRIS services that operate at both the DMC and
Published by Copernicus Publications on behalf of the European Geosciences Union.
32 T. Ahern et al.: New Developments at the IRIS DMC
Figure 1. The new MUSTANG system not only identifies potential problems with data quality but also builds in a feedback loop between
MUSTANG and network operators. This feedback will sometimes be able to help resolve the problem that caused the identified problem and
later MUSTANG estimates should verify the improvement in the data quality.
the ADC include the BUD real time system, the ring server
that replicates files across a computer network and so in prin-
ciple data ingestion from any data source can take place at
either the DMC or the ADC. However real time data inges-
tion for real time streams takes place primarily at the DMC
and would have to be switched to the ADC manually in the
event of a failure. Support for access tools such as BreqFast,
WILBER3, SeismiQuery, WebRequest, and the new MUS-
TANG system all operate at both the DMC and the ADC. In
principle it would be possible to install additional ADCs in
the US or around the world if resources were available.
2 Enhanced quality assessment of time series data
At the current time a very ambitious project called MUS-
TANG is soon to enter operational status at the IRIS DMC
and the ADC. As new time series arrive at IRIS either in
real time or by delayed file transfer procedures, a system of
roughly 50 metric calculators derives statistical metrics that
characterize a day’s worth of waveform data. Such things as
gaps per day, overlaps per day, means, RMS values, medians,
extreme values and other statistical measures are estimated
by MUSTANG algorithms. Additional metrics that are cal-
culated include measures of latency, power spectral density,
power density functions and a variety of other time series
comparisons between multiple time series (collocated sen-
sors, comparisons to nearby stations, comparisons to syn-
thetic tide estimates are calculated) as well as the signal
to noise ratio for windows of data recorded during larger
events. The entire list of metrics being calculated can be
found at http://service.iris.edu/mustangbeta/metrics/1/query
and the list will evolve dynamically as new ways of looking
at data quality are determined with time.
Figure 1 shows the basic concept of MUSTANG. Time
series enter the DMC, metric calculators are run, normally
about one day after real time, and all relevant metrics are
estimated from the new data. These metrics are stored in
a PostgreSQL database and are made accessible through a
set of web services similar to the other data access web ser-
vices that are available at IRIS and some other FDSN cen-
ters. Data technicians are alerted to patterns in a single metric
or a combination of multiple metrics that are indicative of a
data problem. The technicians will validate the data problem,
try to identify the source of the problem and then communi-
cate with knowledgeable people at the seismic network from
which the data came.
At the current time (November 2014) MUSTANG is in
beta mode but appears stable and will enter a production
phase early in 2015. The system is presently being used
to calculate all the metrics for all the relevant data in the
archive. By the end of 2014 we anticipate that all metrics will
have been calculated for IRIS generated data (_GSN, _PASS-
CAL, _OBSIP, _US-Array) as well as significant portions of
the data from networks that share their data with IRIS. A
coverage service will be available that allows one to quickly
assess whether or not metrics have been calculated for a spe-
cific network, station, channel for a given time period.
A sophisticated recalculation component of MUSTANG
is being developed to know when to recalculate metrics. This
can happen when any of the following occur: (1) a new ver-
sion of the time series is received, (2) relevant metadata is
updated, or (3) the algorithm itself changes. When completed
the automated recalculation engine will trigger recalculation
of just those metrics that need recalculation. When complete,
metrics should not become stale for any reason and users will
have confidence that the metrics they view are correct and
represent the metadata and waveform state in the holdings of
the DMC. Ultimately we intend to make use of the various
metrics to enable data requestors to filter the data they re-
ceive from a request to the IRIS DMC based on the values of
the MUSTANG metrics.
Adv. Geosci., 40, 31–35, 2015 www.adv-geosci.net/40/31/2015/
T. Ahern et al.: New Developments at the IRIS DMC 33
Figure 2. The MUSTANG Data Browser allows visualization of
metrics. For instance Box Plots for an entire network can be dis-
played and quickly allow the operator of a specific seismic net-
work to identify specific stations that have problems indicated by
any specific metric. This example shows a boxplot for three sta-
tions of the IRIS/USGS network FDSN Network Code IU. The box
visible for the topmost station shows the 25th, 50th (median) and
75th percentile range of the gaps per day metric. By looking at long
time spans for entire networks in one display, a network analyst
can quickly identify problematic stations within their network. The
small circles show outliers for the given station.
3 Federation of seismological data centers
Seismological activities have been coordinated globally
since the late 1980’s by the FDSN (www.fdsn.org). Driven
be coordinated definitions of web services and standardized
XML schema (Casey and Ahern, 2011), in Working Groups
II and III of the FDSN, identical services have now been
deployed at 3 data centers in the United States and 6 cen-
ters in Europe. In the US, the participating data centers in-
clude (IRIS, the Northern California Earthquake Data Center
(NCEDC), and the National Earthquake Information Center
(NEIC) of the US Geological Survey (USGS). In Europe the
participating centers include the French National Data Center
(RESIF), the Swiss Seismological Service in Zurich (ETHZ),
the Italian National Center for Geophysics and Volcanology
(INGV), the ORFEUS Data Centre in the Netherlands, the
GEOFON data center in Germany, and the International Seis-
mological Centre in the UK.
Each of these centers has exposed FDSN standard web
services that accept identical parameters as a query string
in the URL as well as delivering the same FDSN approved
XML document resulting from the query. These XML defi-
nitions include StationXML for metadata about seismic sta-
tions and channels and QuakeML for returning catalogs of
earthquakes and other seismic events. Waveform data are re-
turned through identical services in miniseed format defined
by the FDSN or as a variety of other formats that include
picture files, sound files, or ASCII files. Waveforms are not
returned as XML.
Figure 3. The MUSTANG Data Browser can also display values of
metrics for arbitrary lengths of time as requested by a user. The pic-
ture above shows the latency for the vertical channel from the GSN
station in Albuquerque, New Mexico, USA. The latency is shown
for a time range of 14 months in this figure. In the new MUSTANG
system latency is estimated once every 4h that allows most latency
problems to be detected and when possible, corrected.
Figure 4. Probability Density Function (PDF) Plots (McNamara
and Buland, 2004). The new MUSTANG Data Browser continu-
ously generates estimates of the power spectral density functions.
From these raw values, PDFs can be generated for specific network-
station-channels for arbitrary time spans. PDFs are an extremely
powerful tool that characterizes the noise across a broad range of
frequencies for a given seismic station. When compared to the Low
Noise Model (shown by the bottommost grey line in the above fig-
ure) a stations performance can easily be ascertained.
The key to the federated services is that other than the left-
most portion of the URL (that points to a specific center), the
right portion of the URL is identical and the resulting XML
document or miniseed data are also identical.
This simplifies the manner in which an external user can
interact with all of the federated centers.
IRIS is also developing a federator. The federator works
as follows. On a routine basis the IRIS federator queries the
www.adv-geosci.net/40/31/2015/ Adv. Geosci., 40, 31–35, 2015
34 T. Ahern et al.: New Developments at the IRIS DMC
Figure 5. These maps show the various FDSN services that have
been placed into operation at various data centers in the US and
Europe. The red diamond indicates that time series services are in
place; blue indicates that station metadata services are in operation,
and yellow means that FDSN event services are in operation.
holdings of all known federated centers to determine the var-
ious time series holdings at each center. It databases this in-
formation in a PostgreSQL database at the DMC. An external
client that could be written in a variety of languages can for-
mulate a request for data in a specific region, bounding box,
or within a specific distance of a point. The IRIS federator
returns information from which specific URLs to retrieve the
waveform and metadata directly from various federated data
centers that hold the information.
4 EarthCube and cross disciplinary data integration
As part of an EarthCube Building Block project, IRIS is see-
ing if the simple web services concepts developed within
seismology can be extended to other geoscience communi-
ties. Driven by a use case in geodynamics, and working with
multiple partners in the US our goal is to ease data discov-
ery, data access, and data usability across several fields in
the earth sciences. The Geoscience Web Services (GeoWS)
building block, advocates for standard approaches in the de-
velopment of simple web services that include standardized
parameter naming conventions, URL usages, similar docu-
mentation styles, and availability of URL builders to show
how URLs to access services can be properly formed, Our
funded partners include Caltech (GPLATES), Columbia Uni-
versity (marine geoscience data), IRIS (seismology), SDSC
(hydrology), UNAVCO (geodesy), and UNIDATA (atmo-
spheric sciences). These six groups will expose their data
holdings through similar style web services. Simultaneously,
IRIS is working directly with other groups in geoscience
to expose their data holdings through simple web services.
These include (1) superconducting gravimeter data, (2) grav-
ity data, magnetic data, structural geology data sets, vol-
canological data and data from three other large facilities in-
cluding the National Geophysical Data Center, Ocean Obser-
vatory Initiative (OOI), and NEON the National Ecological
Observatory Network.
Figure 6. The IRIS Federator Catalog Web Service. The IRIS fed-
erator catalog is a web service that enables a client application to
recover data of interest from the federated system. Periodically, the
IRIS federator queries (red arrows) all of the known external data
centers (small yellow circles) and stores the state of data holdings of
the federated system in a PostgreSQL database. Using the federator
catalog web service to perform client-side federation is a 2-step pro-
cess: (1) a user client submits a request to the catalog service (blue
arrow) to extract a list of data centers and data that match the query
and (2) the user client submits the requests directly to the identified
data centers. The federator can be queried in a manner whereby all
instances of a seismogram can be returned or it can use a powerful
set of business rules to return data from the “authoritative” center
only. If this system proves effective in meeting data user’s needs it
will solve many aspects of accessing the data wanted when those
data are managed in multiple centers around the world.
While we are certain progress will be made we are also
aware that it is impossible to make interdisciplinary data
seamlessly accessible across these 14 domains. For this rea-
son we are working closely with the Global Earth Observa-
tion System of Systems (GEOSS) brokering group that will
offer mediation services across a subset of these domains. If
successful it will ease the task of integration of data from
these 14 domains.
Acknowledgements. Much of the work involved in these projects
drew heavily upon the work of the FDSN working groups es-
pecially Working Groups II and III. In addition to WG II chair
Reinoud Sleeman of ORFEUS, we would like to particularly thank
Marcelo Bianci at GFZ as well as Luca Trani and Alessandro Spi-
nuso at ORFEUS for their very active involvement and timely com-
ments. We would also like to acknowledge the support received by
Steve Bohlen, Bill Walter, and Jennifer Aquillino of Lawrence Liv-
ermore National Laboratory of the US Department of Energy for
helping with the details of establishing the ADC at LLNL.
The developments presented in this paper were supported by
several grants from the National Science Foundation including
EAR-1261681 (SAGE), ICER-1343709 (EarthCube), and ICER-
1321600 (COOPEUS)
Edited by: D. Pesaresi
Reviewed by: two anonymous referees
Adv. Geosci., 40, 31–35, 2015 www.adv-geosci.net/40/31/2015/
T. Ahern et al.: New Developments at the IRIS DMC 35
References
Ahern, T. K.: The FDSN and IRIS Data Management System:
Providing easy access to terabytes of information, Interna-
tional Handbook of Earthquake and Engineering Seismology, 81,
1645–1655, 2003.
Casey, R. and Ahern, T.: Web Services for Seismic Data Archives,
Geoinformatics: Cyberinfrastructure for the Solid Earth Sci-
ences, Cambridge University Press, Part V, 13, 210–223, 2011.
McNamara, D. E. and Buland, R. P.: Ambient Noise Levels in the
Continental United States, B. Seismol. Soc. Am., 94, 1517–1527,
2004.
www.adv-geosci.net/40/31/2015/ Adv. Geosci., 40, 31–35, 2015
... The challenges in maintaining an extended and diverse network of geophysical instruments are to verify that (a) the configuration of each instrument in the field is consistent with its metadata, and (b) the performance of the instruments does not degrade over time. Developments toward (automated) geophysical data quality monitoring systems have been made over the last decades, for example, PQLX (McNamara & Boaz, 2006a), MUSTANG (Ahern et al., 2015), DQA (Ringler et al., 2015), WFCatalog (Trani et al., 2017), AutoStatsQ (Petersen et al., 2019), and component ratio monitoring (Pedersen et al., 2020). These systems are often designed to detect and identify data anomalies due to changes in, for example, local site conditions, technical instrumental problems, timing issues, transmission failures, and to reveal inconsistencies in the instrument metadata. ...
... In the seismic band, above the resonant frequency, it is necessary to correct data for the instrument frequency response when comparing data with ground motion models. The recommendation of Francis et al. (2011) and others is emphasized, for operators to determine the full bandwidth frequency response of the instrument, similar to, for example, the Network Of Superconducting Gravimeters (1997), and publish the transfer function in widely used poles and zeros formats, for example, StationXML (Ahern et al., 2015). ...
Article
Full-text available
This study describes an automated data quality verification procedure supported by a database of power spectral densities (PSD) estimates for geophysical waveform data. The Royal Netherlands Meteorological Institute (KNMI) manages a 100‐TB archive of continuous geophysical data collected from accelerometers, geophones, broadband seismometers, and infrasonic arrays deployed across the continental and Caribbean Netherlands. This rapidly expanding network at a scale of over 700 instruments makes the manual evaluation of data quality impractical and must be supported by data policies and automated methods. A technique is presented to compress and store PSD estimates in a database with a storage footprint of less than 0.05% of the raw data archive. Every week, the instrument performance is validated by comparing statistical properties of its latest monthly probabilistic PSD distribution to strict quality metrics. The criteria include thresholds based on global noise models, datalogger quantization noise models, constraints imposed by ambient noise conditions, and confidence intervals based on PSD estimates calculated from validated archived data. When a threshold is crossed, the station operator is alerted of the suspected degraded instrument performance, severely limiting the required amount of manual labor and associated human errors. The automated PSD assessment technique is applicable to accelerometers, geophones, broadband seismometers, infrasonic stations, and is demonstrated to be extendable to hydrophones, gravimeters, tiltmeters, and Global Navigation Satellite System receivers. The approach is therefore suitable for other geophysical monitoring infrastructures, for example, observational networks dedicated to continuous volcano monitoring. It is shown that it possible to detect degraded instrument performance that may otherwise remain undetected.
... Active assessments that involve station maintenance visits are costly, time-consuming, and require perpetual planning and effort. Methods for passive quality assessment are often pursued due to their advantages in terms of scalability and reduced cost (McNamara and Boaz, 2006;Ahern et al., 2015;Ringler et al., 2015;Trani et al., 2017;Petersen et al., 2019;Pedersen et al., 2020;Koymans et al., 2021). Moreover, such passive techniques do not disturb the measurement setup itself and may be useful in, e.g., citizen science (Raspberry Shake, S.A., 2016) where the acquisition of high quality data can not be guaranteed. ...
Article
Full-text available
The electrical network frequency (ENF) of the alternating current operated on the power grid is a well-known source of noise in digital recordings. The noise is widespread and appears not just in close proximity to high-voltage power lines, but also in instruments simply connected to the mains powers grid. This omnipresent, anthropogenic signal is generally perceived as a nuisance in the processing of geophysical data. Research has therefore been mainly focused on its elimination from data, while its benefits have gone largely unexplored. It is shown that mHz fluctuations in the nominal ENF (50-60 Hz) induced by variations in power usage can be accurately extracted from geophysical data. This information represents a persistent time-calibration signal that is coherent between instruments over national scales. Cross-correlation of reliable reference ENF data published by electrical grid operators with estimated ENF data from geophysical recordings allows timing errors to be resolved at the 1 s level. Furthermore, it is shown that a polarization analysis of particle motion at the ENF can detect instrument orientation anomalies. While the source of the ENF signal in geophysical data appears instrument and site specific, its general utility in the detection of timing and orientation anomalies is presented.
... Distributing data across several network sites partially addresses the problem because the data can be found and accessed faster if there is more than one source [Ahern et al., 2014]. To achieve this, data is replicated among several nodes in the network, and special software is utilized to en-BitTorrent is Apt for Geophysical Data. . . ...
Article
This article covers a nouveau idea of a way to collect and handle geophysical data with a peer-to-peer network in near real-time. The text covers a brief introduction to the cause, the technology, and the particular case of collecting data from Global Navigation Satellite System (GNSS) stations. We describe the proof-of-concept implementation that has been tested. The test was conducted with an experimental GNSS station and a data aggregation facility. In the test, original raw GNSS signal measurements were transferred to the data aggregation center and subsequently to the consumer. Our implementation utilized the BitTorrent protocol to communicate and transfer data. The solution could be used to enable the majority of data aggregation centers for fast, reliable, and transparent real-time data handling experience for the benefit of the scientific community.
... Distributing data across several network sites partially addresses the problem because the data can be found and accessed faster if there is more than one source [5]. To achieve this, data is replicated among several nodes in the network, and special software is utilized to enable it. ...
Preprint
Full-text available
This article covers a nouveau idea of how to collect and handle geophysical data with a peer-to-peer network in near real-time. The text covers a brief introduction to the cause, the technology, and the particular case of collecting data from GNSS stations. We describe the proof-of-concept implementation that has been tested. The test was conducted with an experimental GNSS station and a data aggregation facility. In the test, original raw GNSS signal measurements were transferred to the data aggregation center and subsequently to the consumer. Our implementation utilized BitTorrent to communicate and transfer data. The solution could be used to establish the majority of data aggregation centers activities to provide fast, reliable, and transparent real-time data handling experience to the scientific community.
... The quality assessment of real-time seismic data has become the most critical part of monitoring and forecasting, and the quality of seismic data has a direct impact on the accuracy of the results of the earthquake reporting and early warning system. Incorporated Research Institutions for Seismology (IRIS) [1]currently has hundreds of seismic data producing institutions around the world, The IRIS Data Management Center(DMS) [2] was established in 1986 in Seattle, USA, to provide data management, such as data processing, member sharing. It was also one of the first institutions to raise the issue of seismic data quality [3], but it has been a long process from raising the issue of data quality to quantitatively solving the data quality analysis tools. ...
Article
Full-text available
The quality of real-time waveform data from seismic networks not only directly affects the accuracy of seismic positioning accuracy, but also has an irreversible impact on the credibility of in-depth research findings in seismic disciplines. The article introduces an engineering perspective evaluation method for seismic data quality assessment, and uses real-time waveform data from the seismic station network in Shandong to form a visualization report on the quality of seismic quasi-real-time data, so as to determine the data quality of the seismic station network more intuitively and accurately.
... 2010), the disturbance of ground materials generates crushing stress, and further generates gravitational acceleration and high forward pulse. Zhou Baofeng (2012) held that, such asymmetric waveform may be caused by faults of the strong-motion seismograph [11][12][13][14][15][16][17]. As can be seen in one of the aftershock records of Lushan Earthquake, the original acceleration waveform had its baseline shift upwards around 50s. Similar phenomena were also observed in the aftershock records of Wenchuan Earthquake in 2008. ...
Article
Full-text available
The quality of strong-motion observation data can directly influence the reliability of earthquake disaster prevention, seismic intensity rapid reporting, earthquake early warning and other results, so it’s of great significance to analyse and evaluate the quality of observation data. In this paper, a detailed survey is made considering the status of domestic and overseas studies on the quality of strong-motion observation data, followed by elaborations about the method and deficiencies in the quality evaluation of strong-motion data; Meanwhile, the main factors influencing the quality of strong-motion observation data are also illustrated, and indexes of quality evaluation are summarized by combining with characteristics and application objectives of strong earthquake observation data; Furthermore, this paper also provides discussions about the development trend of data quality evaluation based on the actual strong-motion observation conditions in China, and puts forward some suggestions.
... The dashed grey lines indicate the Peterson high and low noise models. Spectra were acquired from the Modular Utility for STAtistical kNowledge Gathering (MUSTANG) system (Ahern et al. 2015). The spectra are shown for the (a) vertical component (HHZ), (b) horizontal component 1 (HH1) and (c) horizontal component 2 (HH2). ...
... In general, big data refers to can't within tolerable time with traditional information technology and the hardware and software tools to perceive, acquisition, management, data collection, processing and services [3]. According to the characteristics of the electric power industry, electric power big data mainly comes from electric power production, management, operation, smart grid, the three areas. ...
Article
In seismic data, the published metadata are usually used to record instrument responses and their accuracies, which include zero-frequency sensitivity for seismometer, data logger, and other frequency-dependent elements. In this work, we focused on the variations of the seismometer sensitivity with time at the Incorporated Research Institutions for Seismology/International Deployment of Accelerometers (IRIS/IDA) Global Seismographic Network (GSN) seismic stations, with the goal of improving the metadata accuracy for the history of the network. There are several ways to measure the accuracy of seismometer sensitivity for the seismic stations in service. An effective practice recently adopted by the GSN is to collocate a reference seismometer in proximity to permanently installed sensors to verify the in situ sensors' calibration. For those stations with a secondary seismometer, the IRIS MUSTANG system introduced a transfer function metric to reflect two sensors' gain ratios in the microseism frequency band, 4-30 s. In addition, a simulation approach based on measurements of the M2 tidal line (44,712 s period) has proven effective. In this work, we compare and analyze the results from three different sensitivity estimation methods and conclude that the collocated-sensor method is the most stable and reliable with the minimum uncertainties for the entire network life. However, for calibration epochs lacking both the collocated sensor and secondary seismometer, we rely on the analysis results from the tidal method. For IRIS/IDA data since 1992, we obtained over 1100 revised seismometer sensitivity values spanning all epochs.
Article
Receiver functions are a versatile tool commonly used in seismic imaging. Depending on how they are processed, they can be used to image discontinuity structure within the crust or mantle or they can be inverted for seismic velocity either directly or jointly with complementary datasets. However, modern studies generally require large datasets which can be challenging to handle; therefore, FuncLab was originally written as an interactive Matlab GUI to assist in handling these large datasets. This software uses a project database to allow interactive trace editing, data visualization, H-κ stacking for crustal thickness and Vp/Vs ratio, and common conversion point stacking while minimizing computational costs. Since its initial release, significant advances have been made in the implementation of web services and changes in the underlying Matlab platform have necessitated a significant revision to the software. Here, we present revisions to the software, including new features such as data downloading via irisFetch.m, receiver function calculations via processRFmatlab, on-the-fly cross-section tools, interface picking, and more. In the descriptions of the tools, we present its application to a test dataset in Michigan, Wisconsin, and neighboring areas following the passage of USArray Transportable Array. The software is made available online at https://robporritt.wordpress.com/software and the IRIS seismic software repository, https://seiscode.iris.washington.edu/.
Article
Full-text available
We present a new approach to characterize the background seismic noise across the continental United States. Using this approach, power spectral den-sity (PSD) is estimated at broadband seismic stations for frequencies ranging from 0.01 to 16 Hz. We selected a large number of 1-hr waveform segments during a 3-yr period, from 2001 to 2003, from continuous data collected by the U.S. National Seismograph Network and the Advanced National Seismic System (ANSS). For each segment of continuous data, the PSD is estimated and smoothed in full-octave averages at 1/8 octave intervals. Powers for each 1/8 period interval were then accumulated in 1-dB power bins. A statistical analysis of power bins yields probability density functions (PDFs) as a function of noise power for each of the octave bands at each station and component. There is no need to account for earth-quakes since they map into a background probability level. A comparison of day and night PDFs and an examination of artifacts related to station operation and episodic cultural noise allow us to estimate both the overall station quality and the level of earth noise at each site. Percentage points of the PDFs have been derived to form the basis for noise maps of the contiguous United States at body-wave frequencies. The results of our noise analysis are useful for characterizing the performance of existing broadband stations and for detecting operational problems and should be relevant to the future siting of ANSS backbone stations. The noise maps at body-wave frequencies should be useful for estimating the magnitude threshold for the ANSS backbone and regional networks or conversely for optimizing the distribution of regional network stations.
Chapter
Introduction: Seismology has a long tradition of sharing data across boundaries, be they international, institutional, or scientific domain. This sharing began decades before the era of digital data but modern methods of digital access to seismological data have revolutionized the ways in which these data are now available. Normally, seismological information is accessible on a global basis within seconds of real time. Since scientific studies also demand access to data spanning decades, access to voluminous archives of time series data is also necessary. This paper attempts to highlight some of the more recent applications of web services to provide access to seismological data of several types. There are thousands of seismic stations operating in the world today; most are used for monitoring seismicity on a local, national, regional, or global scale. Broadband seismometers are now more and more common and greatly enhance the scientific usefulness of these data. The IRIS Data Management Center (DMC) is a major center that incorporates data from several seismic networks, including those funded by the US National Science Foundation, US Geological Survey, and networks funded by other international and national sources. As an international data center, the IRIS DMC actively works with international groups in a manner where data from many networks around the world, funded by international sources, are available through the IRIS DMC, as well as through distributed data centers.
Article
The amount of data now available from FDSN data centers is vast. The FDSN serves as the only international body that co-ordinates the activities of networks that are deploying seismic stations on both national and international scales. Its work in the development of a standard data exchange format, SEED, and in developing standardized data request tools is very important if the task of data gathering is to remain simple for the average seismological researcher. Although we know that the tools and methods of accessing data will continue to evolve, it is most probable that the amount of data will continue to increase. IRIS and members of the FDSN will continue to make improvements in their systems to ensure that the high-quality data of the FDSN member networks continues to be made available in a timely and easy-to-use manner.
Web Services for Seismic Data Archives, Geoinformatics: Cyberinfrastructure for the Solid Earth Sciences
  • R Casey
  • T Ahern
Casey, R. and Ahern, T.: Web Services for Seismic Data Archives, Geoinformatics: Cyberinfrastructure for the Solid Earth Sciences, Cambridge University Press, Part V, 13, 210-223, 2011.