Experimental Data Set of Mining-Induced Seismicity for Studies of Full-Scale Topographic Effects

2015 ◽  
Vol 31 (1) ◽  
pp. 541-564 ◽  
Author(s):  
Clinton M. Wood ◽  
Brady R. Cox

This paper describes two large, high-quality experimental data sets of ground motions collected with locally dense arrays of seismometers deployed on steep mountainous terrain with varying slope angles and topographic features. These data sets were collected in an area of central-eastern Utah that experiences frequent and predictable mining-induced seismicity as a means to study the effects of topography on small-strain seismic ground motions. The data sets are freely available through the George E. Brown, Jr. Network for Earthquake Engineering Simulation data repository ( NEEShub.org ) under the DOI numbers 10.4231/D34M9199S and 10.4231/D3Z31NN4J. This paper documents the data collection efforts and metadata necessary for utilizing the data sets, as well as the availability of supporting data (e.g., high-resolution digital elevation models). The paper offers a brief summary of analyses conducted on the data sets thus far, in addition to ideas about how these data sets may be used in future studies related to topographic effects and mining seismicity.

2020 ◽  
Vol 224 (1) ◽  
pp. 230-240
Author(s):  
Sean W Johnson ◽  
Derrick J A Chambers ◽  
Michael S Boltz ◽  
Keith D Koper

SUMMARY Monitoring mining-induced seismicity (MIS) can help engineers understand the rock mass response to resource extraction. With a thorough understanding of ongoing geomechanical processes, engineers can operate mines, especially those mines with the propensity for rockbursting, more safely and efficiently. Unfortunately, processing MIS data usually requires significant effort from human analysts, which can result in substantial costs and time commitments. The problem is exacerbated for operations that produce copious amounts of MIS, such as mines with high-stress and/or extraction ratios. Recently, deep learning methods have shown the ability to significantly improve the quality of automated arrival-time picking on earthquake data recorded by regional seismic networks. However, relatively little has been published on applying these techniques to MIS. In this study, we compare the performance of a convolutional neural network (CNN) originally trained to pick arrival times on the Southern California Seismic Network (SCSN) to that of human analysts on coal-mine-related MIS. We perform comparisons on several coal-related MIS data sets recorded at various network scales, sampling rates and mines. We find that the Southern-California-trained CNN does not perform well on any of our data sets without retraining. However, applying the concept of transfer learning, we retrain the SCSN model with relatively little MIS data after which the CNN performs nearly as well as a human analyst. When retrained with data from a single analyst, the analyst-CNN pick time residual variance is lower than the variance observed between human analysts. We also compare the retrained CNN to a simpler, optimized picking algorithm, which falls short of the CNN's performance. We conclude that CNNs can achieve a significant improvement in automated phase picking although some data set-specific training will usually be required. Moreover, initializing training with weights found from other, even very different, data sets can greatly reduce the amount of training data required to achieve a given performance threshold.


Polymers ◽  
2021 ◽  
Vol 13 (21) ◽  
pp. 3811
Author(s):  
Iosif Sorin Fazakas-Anca ◽  
Arina Modrea ◽  
Sorin Vlase

This paper proposes a new method for calculating the monomer reactivity ratios for binary copolymerization based on the terminal model. The original optimization method involves a numerical integration algorithm and an optimization algorithm based on k-nearest neighbour non-parametric regression. The calculation method has been tested on simulated and experimental data sets, at low (<10%), medium (10–35%) and high conversions (>40%), yielding reactivity ratios in a good agreement with the usual methods such as intersection, Fineman–Ross, reverse Fineman–Ross, Kelen–Tüdös, extended Kelen–Tüdös and the error in variable method. The experimental data sets used in this comparative analysis are copolymerization of 2-(N-phthalimido) ethyl acrylate with 1-vinyl-2-pyrolidone for low conversion, copolymerization of isoprene with glycidyl methacrylate for medium conversion and copolymerization of N-isopropylacrylamide with N,N-dimethylacrylamide for high conversion. Also, the possibility to estimate experimental errors from a single experimental data set formed by n experimental data is shown.


2017 ◽  
Author(s):  
Alexander P. Browning ◽  
Scott W. McCue ◽  
Rachelle N. Binny ◽  
Michael J. Plank ◽  
Esha T. Shah ◽  
...  

AbstractCollective cell spreading takes place in spatially continuous environments, yet it is often modelled using discrete lattice-based approaches. Here, we use data from a series of cell proliferation assays, with a prostate cancer cell line, to calibrate a spatially continuous individual based model (IBM) of collective cell migration and proliferation. The IBM explicitly accounts for crowding effects by modifying the rate of movement, direction of movement, and the rate of proliferation by accounting for pair-wise interactions. Taking a Bayesian approach we estimate the free parameters in the IBM using rejection sampling on three separate, independent experimental data sets. Since the posterior distributions for each experiment are similar, we perform simulations with parameters sampled from a new posterior distribution generated by combining the three data sets. To explore the predictive power of the calibrated IBM, we forecast the evolution of a fourth experimental data set. Overall, we show how to calibrate a lattice-free IBM to experimental data, and our work highlights the importance of interactions between individuals. Despite great care taken to distribute cells as uniformly as possible experimentally, we find evidence of significant spatial clustering over short distances, suggesting that standard mean-field models could be inappropriate.


2010 ◽  
Vol 75 (4) ◽  
pp. 483-495 ◽  
Author(s):  
Slavica Eric ◽  
Marko Kalinic ◽  
Aleksandar Popovic ◽  
Halid Makic ◽  
Elvisa Civic ◽  
...  

Aqueous solubility is an important factor influencing several aspects of the pharmacokinetic profile of a drug. Numerous publications present different methodologies for the development of reliable computational models for the prediction of solubility from structure. The quality of such models can be significantly affected by the accuracy of the employed experimental solubility data. In this work, the importance of the accuracy of the experimental solubility data used for model training was investigated. Three data sets were used as training sets - Data Set 1 containing solubility data collected from various literature sources using a few criteria (n = 319), Data Set 2 created by substituting 28 values from Data set 1 with uniformly determined experimental data from one laboratory (n = 319) and Data Set 3 created by including 56 additional components, for which the solubility was also determined under uniform conditions in the same laboratory, in the Data Set 2 (n = 375). The selection of the most significant descriptors was performed by the heuristic method, using one-parameter and multi-parameter analysis. The correlations between the most significant descriptors and solubility were established using multi-linear regression analysis (MLR) for all three investigated data sets. Notable differences were observed between the equations corresponding to different data sets, suggesting that models updated with new experimental data need to be additionally optimized. It was successfully shown that the inclusion of uniform experimental data consistently leads to an improvement in the correlation coefficients. These findings contribute to an emerging consensus that improving the reliability of solubility prediction requires the inclusion of many diverse compounds for which solubility was measured under standardized conditions in the data set.


Author(s):  
Özlem Türkşen ◽  
Suna Ertunç

Beta-glucan (BG) has positive health effects for the mamalians. However, the BG sources have limited content of it. Besides, the production of the BG has stringent procedures with low productivity. Economical production of the BG needs the improvement of the BG production steps. In this study, it is aimed to improve the BG content during the first step of the BG production, microorganism growth step, by obtaining the optimal values of additive materials (EDTA, CaCl2 and Sorbitol). For this purpose, the experimental data sets with replicated response measures (RRM) are obtained at spesific levels of EDTA, CaCl2 and Sorbitol. Fuzzy modeling, a flexible modeling approach, is applied on the experimental data set because of the small sized data set and diffulty of satisfying probabilistic modeling assumptions. The predicted fuzzy function is obtained according to the fuzzy least squares approach. In order to get the optimal values of EDTA, CaCl2 and Sorbitol, the predicted fuzzy function is maximized based on multi-objective optimization (MOO) approach. By using the optimal values of EDTA, CaCl2 and Sorbitol, the uncertainty for predicted BG content is evaluated from the economic perspective.


2021 ◽  
Author(s):  
Wouter Dorigo ◽  
Irene Himmelbauer ◽  
Daniel Aberer ◽  
Lukas Schremmer ◽  
Ivana Petrakovic ◽  
...  

Abstract. In 2009, the International Soil Moisture Network (ISMN) was initiated as a community effort, funded by the European Space Agency, to serve as a centralised data hosting facility for globally available in situ soil moisture measurements (Dorigo et al., 2011a, b). The ISMN brings together in situ soil moisture measurements collected and freely shared by a multitude of organisations, harmonizes them in terms of units and sampling rates, applies advanced quality control, and stores them in a database. Users can freely retrieve the data from this database through an online web portal (https://ismn.earth). Meanwhile, the ISMN has evolved into the primary in situ soil moisture reference database worldwide, as evidenced by more than 3000 active users and over 1000 scientific publications referencing the data sets provided by the network. As of December 2020, the ISMN now contains data of 65 networks and 2678 stations located all over the globe, with a time period spanning from 1952 to present.The number of networks and stations covered by the ISMN is still growing and many of the data sets contained in the database continue to be updated. The main scope of this paper is to inform readers about the evolution of the ISMN over the past decade,including a description of network and data set updates and quality control procedures. A comprehensive review of existing literature making use of ISMN data is also provided in order to identify current limitations in functionality and data usage, and to shape priorities for the next decade of operations of this unique community-based data repository.


2016 ◽  
Vol 72 (6) ◽  
pp. 696-703 ◽  
Author(s):  
Julian Henn

An alternative measure to the goodness of fit (GoF) is developed and applied to experimental data. The alternative goodness of fit squared (aGoFs) demonstrates that the GoF regularly fails to provide evidence for the presence of systematic errors, because certain requirements are not met. These requirements are briefly discussed. It is shown that in many experimental data sets a correlation between the squared residuals and the variance of observed intensities exists. These correlations corrupt the GoF and lead to artificially reduced values in the GoF and in the numerical value of thewR(F2). Remaining systematic errors in the data sets are veiled by this mechanism. In data sets where these correlations do not appear for the entire data set, they often appear for the decile of largest variances of observed intensities. Additionally, statistical errors for the squared goodness of fit, GoFs, and the aGoFs are developed and applied to experimental data. This measure shows how significantly the GoFs and aGoFs deviate from the ideal value one.


Author(s):  
James Simek ◽  
Jed Ludlow ◽  
Phil Tisovec

InLine Inspection (ILI) tools using the magnetic flux leakage (MFL) technique are the most common type used for performing metal loss surveys worldwide. Based upon the very robust and proven magnetic flux leakage technique, these tools have been shown to operate reliably in the extremely harsh environments of transmission pipelines. In addition to metal loss, MFL tools are capable of identifying a broad range of pipeline features. Most MFL surveys to date have used tools employing axially oriented magnetizers, capable of detecting and quantifying many categories of volumetric metal loss features. For certain classes of axially oriented features, MFL tools using axially oriented fields have encountered difficulty in detection and subsequent quantification. To address features in these categories, tools employing circumferential or transversely oriented fields have been designed and placed into service, enabling enhanced detection and sizing for axially oriented features. In most cases, multiple surveys are required, as current tools do not incorporate the ability to collect both data sets concurrently. Applying the magnetic field in an oblique direction will enable detection of axially oriented features and may be used simultaneously with an axially oriented tool. Referencing previous research in adapting circumferential or transverse designs for inline service, the concept of an oblique field magnetizer will be presented. Models developed demonstrating the technique are discussed, shown with experimental data supporting the concept. Efforts involved in the implementation of an oblique magnetizer, including magnetic models for field profiles used to determine magnetizer configurations and sensor locations are presented. Experimental results are provided detailing the response of the system to a full range of metal loss features, supplementing modeling in an effort to determine the effects of variables introduced by magnetic property and velocity induced differences. Included in the experimental data results are extremely narrow axially oriented features, many of which are not detected or identified within the axial data set. Experimental and field verification results for detection accuracies will be described in comparison to an axial field tool.


2010 ◽  
Vol 62 (4) ◽  
pp. 875-882 ◽  
Author(s):  
A. Dembélé ◽  
J.-L. Bertrand-Krajewski ◽  
B. Barillon

Regression models are among the most frequently used models to estimate pollutants event mean concentrations (EMC) in wet weather discharges in urban catchments. Two main questions dealing with the calibration of EMC regression models are investigated: i) the sensitivity of models to the size and the content of data sets used for their calibration, ii) the change of modelling results when models are re-calibrated when data sets grow and change with time when new experimental data are collected. Based on an experimental data set of 64 rain events monitored in a densely urbanised catchment, four TSS EMC regression models (two log-linear and two linear models) with two or three explanatory variables have been derived and analysed. Model calibration with the iterative re-weighted least squares method is less sensitive and leads to more robust results than the ordinary least squares method. Three calibration options have been investigated: two options accounting for the chronological order of the observations, one option using random samples of events from the whole available data set. Results obtained with the best performing non linear model clearly indicate that the model is highly sensitive to the size and the content of the data set used for its calibration.


2006 ◽  
Vol 17 (09) ◽  
pp. 1313-1325 ◽  
Author(s):  
NIKITA A. SAKHANENKO ◽  
GEORGE F. LUGER ◽  
HANNA E. MAKARUK ◽  
JOYSREE B. AUBREY ◽  
DAVID B. HOLTKAMP

This paper considers a set of shock physics experiments that investigate how materials respond to the extremes of deformation, pressure, and temperature when exposed to shock waves. Due to the complexity and the cost of these tests, the available experimental data set is often very sparse. A support vector machine (SVM) technique for regression is used for data estimation of velocity measurements from the underlying experiments. Because of good generalization performance, the SVM method successfully interpolates the experimental data. The analysis of the resulting velocity surface provides more information on the physical phenomena of the experiment. Additionally, the estimated data can be used to identify outlier data sets, as well as to increase the understanding of the other data from the experiment.


Sign in / Sign up

Export Citation Format

Share Document