Data-Driven Spatiotemporal Modeling of Reservoir Pressure Dynamics in a Giant Saturated Oil Field in Libya

2021 ◽  
Author(s):  
Saumitra Dwivedi ◽  
Guillaume Suzanne ◽  
Abdulhakim Algadban ◽  
Ibrahim A. Hameed

Abstract This paper aims to explore modern techniques based on artificial intelligence (AI) and data science, in order to produce data-driven workflows to analyze, model, and simulate reservoir pressure dynamics. In this paper, it was investigated a data-driven workflow to model reservoir pressure at any point in space and time from sparse pressure data observed at wells, without building a physics-based numerical model. This workflow was termed as spatiotemporal modelling of reservoir pressure. Spatiotemporal modelling of reservoir pressure was based on a three-step workflow including multivariate analysis of pressure data and relevant explanatory variables (features), pressure modelling and spatiotemporal interpolation. The overall workflow provided a comprehensive method to understand and map the reservoir pressure dynamics using data science tools. Several modelling techniques such as generalized additive models, artificial neural networks and spatiotemporal kriging were investigated for their applicability and accuracy. The workflow was applied to a real oil and gas reservoir case, for which the reservoir pressure prediction accuracy was optimized through a few experiments. The optimum experiment produced highly accurate prediction with a mean absolute error of 26.85 psi measured on the training dataset. Moreover, a portion of data used was kept to evaluate blind test accuracy, which amounted to a mean absolute error of 55 psi, for the optimum case. The proposed data-driven workflow was aimed to improve current methods of reservoir engineering and simulation. The suggested workflow showed high accuracy in reservoir pressure predictions with high efficiency in terms of computational resources and time. Additionally, the proposed workflow was developed using open-source libraries which pose no additional cost to computation, in contrast to extremely expensive industry standard physics-based reservoir simulation software. Finally, this workflow could also be used to model other reservoir variables such as production ratios (Water cut, and Gas-Oil Ratio), contacts (Water-Oil contact and Gas-Oil contact), among others.

Author(s):  
A. R. Nemati ◽  
M. Zakeri Niri ◽  
S. Moazami

Simulation of rainfall-runoff process is one of the most important research fields in hydrology and water resources. Generally, the models used in this section are divided into two conceptual and data-driven categories. In this study, a conceptual model and two data-driven models have been used to simulate rainfall-runoff process in Tamer sub-catchment located in Gorganroud watershed in Iran. The conceptual model used is HEC-HMS, and data-driven models are neural network model of multi-layer Perceptron (MLP) and support vector regression (SVR). In addition to simulation of rainfall-runoff process using the recorded land precipitation, the performance of four satellite algorithms of precipitation, that is, CMORPH, PERSIANN, TRMM 3B42 and TRMM 3B42RT were studied. In simulation of rainfall-runoff process, calibration and accuracy of the models were done based on satellite data. The results of the research based on three criteria of correlation coefficient (R), root mean square error (RMSE) and mean absolute error (MAE) showed that in this part the two models of SVR and MLP could perform the simulation of runoff in a relatively appropriate way, but in simulation of the maximum values of the flow, the error of models increased.


2021 ◽  
Vol 3 (1) ◽  
Author(s):  
Ryan F McCormick ◽  
Sandra K Truong ◽  
Jose Rotundo ◽  
Adam P Gaspar ◽  
Don Kyle ◽  
...  

Abstract ABSTRACTThe timing of crop development has significant impacts on management decisions and subsequent yield formation. A large intercontinental dataset recording the timing of soybean developmental stages was used to establish ensembling approaches that leverage both knowledge-based, human-defined models of soybean phenology and data-driven, machine-learned models to achieve accurate and interpretable predictions. We demonstrate that the knowledge-based models can improve machine learning by generating expert-engineered features. The collection of knowledge-based and data-driven models was combined via super learning to both improve prediction and identify the most performant models. Stacking the predictions of the component models resulted in a mean absolute error of 4.41 and 5.27 days to flowering (R1) and physiological maturity (R7), providing an improvement relative to the benchmark knowledge-based model error of 6.94 and 15.53 days, respectively, in cross-validation. The hybrid intercontinental model applies to a much wider range of management and temperature conditions than previous mechanistic models, enabling improved decision support as alternative cropping systems arise, farm sizes increase and changes in the global climate continue to accelerate.


2020 ◽  
Vol 15 ◽  
Author(s):  
Fahad Layth Malallah ◽  
Baraa T. Shareef ◽  
Mustafah Ghanem Saeed ◽  
Khaled N. Yasen

Aims: Normally, the temperature increase of individuals leads to the possibility of getting a type of disease, which might be risky to other people such as coronavirus. Traditional techniques for tracking core-temperature require body contact either by oral, rectum, axillary, or tympanic, which are unfortunately considered intrusive in nature as well as causes of contagion. Therefore, sensing human core-temperature non-intrusively and remotely is the objective of this research. Background: Nowadays, increasing level of medical sectors is a necessary targets for the research operations, especially with the development of the integrated circuit, sensors and cameras that made the normal life easier. Methods: The solution is by proposing an embedded system consisting of the Arduino microcontroller, which is trained with a model of Mean Absolute Error (MAE) analysis for predicting Contactless Core-Temperature (CCT), which is the real body temperature. Results: The Arduino is connected to an Infrared-Thermal sensor named MLX90614 as input signal, and connected to the LCD to display the CCT. To evaluate the proposed system, experiments are conducted by participating 31-subject sensing contactless temperature from the three face sub-regions: forehead, nose, and cheek. Conclusion: Experimental results approved that CCT can be measured remotely depending on the human face, in which the forehead region is better to be dependent, rather than nose and cheek regions for CCT measurement due to the smallest


Sensors ◽  
2021 ◽  
Vol 21 (8) ◽  
pp. 2670
Author(s):  
Thomas Quirin ◽  
Corentin Féry ◽  
Dorian Vogel ◽  
Céline Vergne ◽  
Mathieu Sarracanie ◽  
...  

This paper presents a tracking system using magnetometers, possibly integrable in a deep brain stimulation (DBS) electrode. DBS is a treatment for movement disorders where the position of the implant is of prime importance. Positioning challenges during the surgery could be addressed thanks to a magnetic tracking. The system proposed in this paper, complementary to existing procedures, has been designed to bridge preoperative clinical imaging with DBS surgery, allowing the surgeon to increase his/her control on the implantation trajectory. Here the magnetic source required for tracking consists of three coils, and is experimentally mapped. This mapping has been performed with an in-house three-dimensional magnetic camera. The system demonstrates how magnetometers integrated directly at the tip of a DBS electrode, might improve treatment by monitoring the position during and after the surgery. The three-dimensional operation without line of sight has been demonstrated using a reference obtained with magnetic resonance imaging (MRI) of a simplified brain model. We observed experimentally a mean absolute error of 1.35 mm and an Euclidean error of 3.07 mm. Several areas of improvement to target errors below 1 mm are also discussed.


Sensors ◽  
2021 ◽  
Vol 21 (11) ◽  
pp. 3719
Author(s):  
Aoxin Ni ◽  
Arian Azarang ◽  
Nasser Kehtarnavaz

The interest in contactless or remote heart rate measurement has been steadily growing in healthcare and sports applications. Contactless methods involve the utilization of a video camera and image processing algorithms. Recently, deep learning methods have been used to improve the performance of conventional contactless methods for heart rate measurement. After providing a review of the related literature, a comparison of the deep learning methods whose codes are publicly available is conducted in this paper. The public domain UBFC dataset is used to compare the performance of these deep learning methods for heart rate measurement. The results obtained show that the deep learning method PhysNet generates the best heart rate measurement outcome among these methods, with a mean absolute error value of 2.57 beats per minute and a mean square error value of 7.56 beats per minute.


Vibration ◽  
2021 ◽  
Vol 4 (2) ◽  
pp. 341-356
Author(s):  
Jessada Sresakoolchai ◽  
Sakdirat Kaewunruen

Various techniques have been developed to detect railway defects. One of the popular techniques is machine learning. This unprecedented study applies deep learning, which is a branch of machine learning techniques, to detect and evaluate the severity of rail combined defects. The combined defects in the study are settlement and dipped joint. Features used to detect and evaluate the severity of combined defects are axle box accelerations simulated using a verified rolling stock dynamic behavior simulation called D-Track. A total of 1650 simulations are run to generate numerical data. Deep learning techniques used in the study are deep neural network (DNN), convolutional neural network (CNN), and recurrent neural network (RNN). Simulated data are used in two ways: simplified data and raw data. Simplified data are used to develop the DNN model, while raw data are used to develop the CNN and RNN model. For simplified data, features are extracted from raw data, which are the weight of rolling stock, the speed of rolling stock, and three peak and bottom accelerations from two wheels of rolling stock. In total, there are 14 features used as simplified data for developing the DNN model. For raw data, time-domain accelerations are used directly to develop the CNN and RNN models without processing and data extraction. Hyperparameter tuning is performed to ensure that the performance of each model is optimized. Grid search is used for performing hyperparameter tuning. To detect the combined defects, the study proposes two approaches. The first approach uses one model to detect settlement and dipped joint, and the second approach uses two models to detect settlement and dipped joint separately. The results show that the CNN models of both approaches provide the same accuracy of 99%, so one model is good enough to detect settlement and dipped joint. To evaluate the severity of the combined defects, the study applies classification and regression concepts. Classification is used to evaluate the severity by categorizing defects into light, medium, and severe classes, and regression is used to estimate the size of defects. From the study, the CNN model is suitable for evaluating dipped joint severity with an accuracy of 84% and mean absolute error (MAE) of 1.25 mm, and the RNN model is suitable for evaluating settlement severity with an accuracy of 99% and mean absolute error (MAE) of 1.58 mm.


2021 ◽  
pp. 1-13
Author(s):  
Richa ◽  
Punam Bedi

Recommender System (RS) is an information filtering approach that helps the overburdened user with information in his decision making process and suggests items which might be interesting to him. While presenting recommendation to the user, accuracy of the presented list is always a concern for the researchers. However, in recent years, the focus has now shifted to include the unexpectedness and novel items in the list along with accuracy of the recommended items. To increase the user acceptance, it is important to provide potentially interesting items which are not so obvious and different from the items that the end user has rated. In this work, we have proposed a model that generates serendipitous item recommendation and also takes care of accuracy as well as the sparsity issues. Literature suggests that there are various components that help to achieve the objective of serendipitous recommendations. In this paper, fuzzy inference based approach is used for the serendipity computation because the definitions of the components overlap. Moreover, to improve the accuracy and sparsity issues in the recommendation process, cross domain and trust based approaches are incorporated. A prototype of the system is developed for the tourism domain and the performance is measured using mean absolute error (MAE), root mean square error (RMSE), unexpectedness, precision, recall and F-measure.


2021 ◽  
pp. 875697282199994
Author(s):  
Joseph F. Hair ◽  
Marko Sarstedt

Most project management research focuses almost exclusively on explanatory analyses. Evaluation of the explanatory power of statistical models is generally based on F-type statistics and the R 2 metric, followed by an assessment of the model parameters (e.g., beta coefficients) in terms of their significance, size, and direction. However, these measures are not indicative of a model’s predictive power, which is central for deriving managerial recommendations. We recommend that project management researchers routinely use additional metrics, such as the mean absolute error or the root mean square error, to accurately quantify their statistical models’ predictive power.


Atmosphere ◽  
2021 ◽  
Vol 12 (7) ◽  
pp. 830
Author(s):  
William E. Lewis ◽  
Timothy L. Olander ◽  
Christopher S. Velden ◽  
Christopher Rozoff ◽  
Stefano Alessandrini

Accurate, reliable estimates of tropical cyclone (TC) intensity are a crucial element in the warning and forecast process worldwide, and for the better part of 50 years, estimates made from geostationary satellite observations have been indispensable to forecasters for this purpose. One such method, the Advanced Dvorak Technique (ADT), was used to develop analog ensemble (AnEn) techniques that provide more precise estimates of TC intensity with instant access to information on the reliability of the estimate. The resulting methods, ADT-AnEn and ADT-based Error Analog Ensemble (ADTE-AnEn), were trained and tested using seventeen years of historical ADT intensity estimates using k-fold cross-validation with 10 folds. Using only two predictors, ADT-estimated current intensity (maximum wind speed) and TC center latitude, both AnEn techniques produced significant reductions in mean absolute error and bias for all TC intensity classes in the North Atlantic and for most intensity classes in the Eastern Pacific. The ADTE-AnEn performed better for extreme intensities in both basins (significantly so in the Eastern Pacific) and will be incorporated in the University of Wisconsin’s Cooperative Institute for Meteorological Satellite Studies (UW-CIMSS) workflow for further testing during operations in 2021.


Sign in / Sign up

Export Citation Format

Share Document