scholarly journals Towards Scalable Gaussian Process Modeling

Author(s):  
Piyush Pandita ◽  
Jesper Kristensen ◽  
Liping Wang

Abstract Numerous engineering problems of interest to the industry are often characterized by expensive black-box objective function evaluations. These objective functions could be physical experiments or computer simulations. Obtaining a comprehensive idea of the problem and/or performing subsequent optimizations generally requires hundreds of thousands of evaluations of the objective function which is most often a practically unachievable task. Gaussian Process (GP) surrogate modeling replaces the expensive function with a cheap-to-evaluate data-driven probabilistic model. While the GP does not assume a functional form of the problem, it is defined by a set of parameters, called hyper-parameters, that need to be learned from the data. The hyperparameters define the characteristics of the objective function, such as smoothness, magnitude, periodicity, etc. Accurately estimating these hyperparameters is a key ingredient in developing a reliable and generalizable surrogate model. Markov chain Monte Carlo (MCMC) is a ubiquitously used Bayesian method to estimate these hyperparameters. At GEs Global Research Center, a customized industry-strength Bayesian hybrid modeling framework utilizing the GP, called GEBHM, has been employed and validated over many years. GEBHM is very effective on problems of small and medium size, typically less than 1000 training points. However, the GP does not scale well in time with a growing dataset and problem dimensionality which can be a major impediment in such problems. For some challenging industry applications, the predictive capability of the GP is required but each second during the training of the GP costs thousands of dollars. In this work, we apply a scalable MCMC-based methodology enabling the modeling of large-scale industry problems. Towards this, we extend and implement in GEBHM an Adaptive Sequential Monte Carlo (ASMC) methodology for training the GP. This implementation saves computational time (especially for large-scale problems) while not sacrificing predictability over the current MCMC implementation. We demonstrate the effectiveness and accuracy of GEBHM with ASMC on four mathematical problems and on two challenging industry applications of varying complexity.

2021 ◽  
pp. 1-11
Author(s):  
Piyush Pandita ◽  
Panagiotis Tsilifis ◽  
Sayan Ghosh ◽  
Liping Wang

Abstract Gaussian Process (GP) regression or kriging has been extensively applied in the engineering literature for the purposes of building a cheap-to-evaluate surrogate, within the contexts of multi-fidelity modeling, model calibration and design optimization. With the ongoing automation of manufacturing and industrial practices as a part of Industry 4.0, there has been greater need for advancing GP regression techniques to handle challenges such as high input dimensionality, data paucity or big data problems, these consist primarily of proposing efficient design of experiments, optimal data acquisition strategies, and other mathematical tricks. In this work, our attention is focused on the challenges of efficiently training a GP model, which, to the authors opinion, has attracted very little attention and is to-date, poorly addressed. The performance of widely used training approaches such as maximum likelihood estimation and Markov Chain Monte Carlo (MCMC) sampling can deteriorate significantly in high dimensional and big data problems and can lead to cost deficient implementations of critical importance to many industrial applications. Here, we compare an Adaptive Sequential Monte Carlo (ASMC) sampling algorithm to classic MCMC sampling strategies and we demonstrate the effectiveness of our implementation on several mathematical problems and challenging industry applications of varying complexity. The computational time savings of our ASMC approach manifest in large-scale problems helping us to push the boundary of GP regression applicability and scalability in various domain of Industry 4.0, including but not limited to design automation, design engineering, predictive maintenance, and supply chain manufacturing.


2000 ◽  
Vol 12 (4) ◽  
pp. 955-993 ◽  
Author(s):  
J. F. G. de Freitas ◽  
M. Niranjan ◽  
A. H. Gee ◽  
A. Doucet

We discuss a novel strategy for training neural networks using sequential Monte Carlo algorithms and propose a new hybrid gradient descent/sampling importance resampling algorithm (HySIR). In terms of computational time and accuracy, the hybrid SIR is a clear improvement over conventional sequential Monte Carlo techniques. The new algorithm may be viewed as a global optimization strategy that allows us to learn the probability distributions of the network weights and outputs in a sequential framework. It is well suited to applications involving on-line, nonlinear, and nongaussian signal processing. We show how the new algorithm outperforms extended Kalman filter training on several problems. In particular, we address the problem of pricing option contracts, traded in financial markets. In this context, we are able to estimate the one-step-ahead probability density functions of the options prices.


2014 ◽  
Vol 501-504 ◽  
pp. 1067-1070
Author(s):  
Li Feng Peng ◽  
Guo Shao Su ◽  
Wei Zhao

The performance function of large-scale complicated engineering structure is always highly nonlinear and implicit, and its reliability needs to be evaluated through a time-consuming Finite Element method (FEM). A new method, Gaussian process classification (GPC) dynamic response surface based on Monte Carlo Simulation (MCS) was proposed. Small training samples were created using FEM and Markov chain. Then, the most probable point (MPP) is predicted quickly using MCS without any extra FEM analysis. Furthermore, an iterative algorithm is presented to reduce the errors of GPC by using information of MPP to improve the reconstructing precision constantly. Then, Monte Carlo method combined with GPC surface is applied to get the probability of failure. Several examples results demonstrate the efficiency and robustness of the proposed method, compared with the results of common reliability methods.


2020 ◽  
pp. 1471082X2091608 ◽  
Author(s):  
Danilo Alvares ◽  
Carmen Armero ◽  
Anabel Forte ◽  
Nicolas Chopin

The statistical analysis of the information generated by medical follow-up is a very important challenge in the field of personalized medicine. As the evolutionary course of a patient's disease progresses, his/her medical follow-up generates more and more information that should be processed immediately in order to review and update his/her prognosis and treatment. Hence, we focus on this update process through sequential inference methods for joint models of longitudinal and time-to-event data from a Bayesian perspective. More specifically, we propose the use of sequential Monte Carlo (SMC) methods for static parameter joint models with the intention of reducing computational time in each update of the full Bayesian inferential process. Our proposal is very general and can be easily applied to most popular joint models approaches. We illustrate the use of the presented sequential methodology in a joint model with competing risk events for a real scenario involving patients on mechanical ventilation in intensive care units (ICUs).


2015 ◽  
Vol 733 ◽  
pp. 156-160
Author(s):  
Xia Yan ◽  
Jun Li ◽  
Hui Zhao

A novel and simple parameterization method using an ensemble of unconditional model realizations is applied to decrease the dimension of the misfit objective function in large-scale history matching problems. The major advantage of this parameterization method is that the singular value decomposition (SVD) calculation is completely avoided, which saves time and cost for huge matrix decomposition and the eigenvectors computations in parameterization process. After objective function transforms from a higher dimension to a lower dimension by parameterization, a Monte Carlo approach is introduced to evaluate the gradient information in the lower domain. Unlike the adjoint-gradient algorithms, the gradient in our method is estimated by Monte Carlo stochastic method, which can be easily coupled with different numerical simulator and avoid complicated adjoint code. When the estimated gradient information is obtained, any gradient-based algorithm can be implemented for optimizing the objective function. The Monte Carlo algorithm combined with the parameterization method is applied to Brugge reservoir field. The result shows that our present method gives a good estimation of reservoir properties and decreases the geological uncertainty without SVD but with a lower final objective function value, which provides a more efficient and useful way for history matching in large scale field.


2021 ◽  
Vol 2 ◽  
Author(s):  
Lawrence M. Murray ◽  
Sumeetpal S. Singh ◽  
Anthony Lee

Abstract Monte Carlo algorithms simulates some prescribed number of samples, taking some random real time to complete the computations necessary. This work considers the converse: to impose a real-time budget on the computation, which results in the number of samples simulated being random. To complicate matters, the real time taken for each simulation may depend on the sample produced, so that the samples themselves are not independent of their number, and a length bias with respect to compute time is apparent. This is especially problematic when a Markov chain Monte Carlo (MCMC) algorithm is used and the final state of the Markov chain—rather than an average over all states—is required, which is the case in parallel tempering implementations of MCMC. The length bias does not diminish with the compute budget in this case. It also occurs in sequential Monte Carlo (SMC) algorithms, which is the focus of this paper. We propose an anytime framework to address the concern, using a continuous-time Markov jump process to study the progress of the computation in real time. We first show that for any MCMC algorithm, the length bias of the final state’s distribution due to the imposed real-time computing budget can be eliminated by using a multiple chain construction. The utility of this construction is then demonstrated on a large-scale SMC $ {}^2 $ implementation, using four billion particles distributed across a cluster of 128 graphics processing units on the Amazon EC2 service. The anytime framework imposes a real-time budget on the MCMC move steps within the SMC $ {}^2 $ algorithm, ensuring that all processors are simultaneously ready for the resampling step, demonstrably reducing idleness to due waiting times and providing substantial control over the total compute budget.


Sign in / Sign up

Export Citation Format

Share Document