Cholesky Factorization on Heterogeneous CPU and GPU Systems

Author(s):  
Jieyang Chen ◽  
Zizhong Chen
2021 ◽  
Vol 402 ◽  
pp. 126037
Author(s):  
Li Chen ◽  
Shuisheng Zhou ◽  
Jiajun Ma ◽  
Mingliang Xu

Author(s):  
Jack Poulson

Determinantal point processes (DPPs) were introduced by Macchi (Macchi 1975 Adv. Appl. Probab. 7 , 83–122) as a model for repulsive (fermionic) particle distributions. But their recent popularization is largely due to their usefulness for encouraging diversity in the final stage of a recommender system (Kulesza & Taskar 2012 Found. Trends Mach. Learn. 5 , 123–286). The standard sampling scheme for finite DPPs is a spectral decomposition followed by an equivalent of a randomly diagonally pivoted Cholesky factorization of an orthogonal projection, which is only applicable to Hermitian kernels and has an expensive set-up cost. Researchers Launay et al. 2018 ( http://arxiv.org/abs/1802.08429 ); Chen & Zhang 2018 NeurIPS ( https://papers.nips.cc/paper/7805-fast-greedy-map-inference-for-determinantal-point-process-to-improve-recommendation-diversity.pdf ) have begun to connect DPP sampling to LDL H factorizations as a means of avoiding the initial spectral decomposition, but existing approaches have only outperformed the spectral decomposition approach in special circumstances, where the number of kept modes is a small percentage of the ground set size. This article proves that trivial modifications of LU and LDL H factorizations yield efficient direct sampling schemes for non-Hermitian and Hermitian DPP kernels, respectively. Furthermore, it is experimentally shown that even dynamically scheduled, shared-memory parallelizations of high-performance dense and sparse-direct factorizations can be trivially modified to yield DPP sampling schemes with essentially identical performance. The software developed as part of this research, Catamari ( hodgestar.com/catamari ) is released under the Mozilla Public License v.2.0. It contains header-only, C++14 plus OpenMP 4.0 implementations of dense and sparse-direct, Hermitian and non-Hermitian DPP samplers. This article is part of a discussion meeting issue ‘Numerical algorithms for high-performance computational science’.


1994 ◽  
Vol 02 (04) ◽  
pp. 371-422 ◽  
Author(s):  
E. PADOVANI ◽  
E. PRIOLO ◽  
G. SERIANI

The finite element method (FEM) is a numerical technique well suited to solving problems of elastic wave propagation in complex geometries and heterogeneous media. The main advantages are that very irregular grids can be used, free surface boundary conditions can be easily taken into account, a good reconstruction is possible of irregular surface topography, and complex geometries, such as curved, dipping and rough interfaces, intrusions, cusps, and holes can be defined. The main drawbacks of the classical approach are the need for a large amount of memory, low computational efficiency, and the possible appearance of spurious effects. In this paper we describe some experience in improving the computational efficiency of a finite element code based on a global approach, and used for seismic modeling in geophysical oil exploration. Results from the use of different methods and models run on a mini-superworkstation APOLLO DN10000 are reported and compared. With Chebyshev spectral elements, great accuracy can be reached with almost no numerical artifacts. Static condensation of the spectral element's internal nodes dramatically reduces memory requirements and CPU time. Time integration performed with the classical implicit Newmark scheme is very accurate but not very efficient. Due to the high sparsity of the matrices, the use of compressed storage is shown to greatly reduce not only memory requirements but also computing time. The operation which most affects the performance is the matrix-by-vector product; an effective programming of this subroutine for the storage technique used is decisive. The conjugate gradient method preconditioned by incomplete Cholesky factorization provides, in general, a good compromise between efficiency and memory requirements. Spectral elements greatly increase its efficiency, since the number of iterations is reduced. The most efficient and accurate method is a hybrid iterative-direct solution of the linear system arising from the static condensation of high order elements. The size of 2D models that can be handled in a reasonable time on this kind of computer is nowadays hardly sufficient, and significant 3D modeling is completely unfeasible. However the introduction of new FEM algorithms coupled with the use of new computer architectures is encouraging for the future.


Author(s):  
Erfan Bank Tavakoli ◽  
Michael Riera ◽  
Masudul Hassan Quraishi ◽  
Fengbo Ren

2014 ◽  
Vol 7 (1) ◽  
pp. 225-241 ◽  
Author(s):  
A. Barth ◽  
J.-M. Beckers ◽  
C. Troupin ◽  
A. Alvera-Azcárate ◽  
L. Vandenbulcke

Abstract. A tool for multidimensional variational analysis (divand) is presented. It allows the interpolation and analysis of observations on curvilinear orthogonal grids in an arbitrary high dimensional space by minimizing a cost function. This cost function penalizes the deviation from the observations, the deviation from a first guess and abruptly varying fields based on a given correlation length (potentially varying in space and time). Additional constraints can be added to this cost function such as an advection constraint which forces the analysed field to align with the ocean current. The method decouples naturally disconnected areas based on topography and topology. This is useful in oceanography where disconnected water masses often have different physical properties. Individual elements of the a priori and a posteriori error covariance matrix can also be computed, in particular expected error variances of the analysis. A multidimensional approach (as opposed to stacking two-dimensional analysis) has the benefit of providing a smooth analysis in all dimensions, although the computational cost is increased. Primal (problem solved in the grid space) and dual formulations (problem solved in the observational space) are implemented using either direct solvers (based on Cholesky factorization) or iterative solvers (conjugate gradient method). In most applications the primal formulation with the direct solver is the fastest, especially if an a posteriori error estimate is needed. However, for correlated observation errors the dual formulation with an iterative solver is more efficient. The method is tested by using pseudo-observations from a global model. The distribution of the observations is based on the position of the Argo floats. The benefit of the three-dimensional analysis (longitude, latitude and time) compared to two-dimensional analysis (longitude and latitude) and the role of the advection constraint are highlighted. The tool divand is free software, and is distributed under the terms of the General Public Licence (GPL) (http://modb.oce.ulg.ac.be/mediawiki/index.php/divand).


Sign in / Sign up

Export Citation Format

Share Document