Gracefully degrading systems using the bulk-synchronous parallel model with randomised shared memory

Author(s):  
A. Savva ◽  
T. Nanya
Author(s):  
Xing Zhao ◽  
Manos Papagelis ◽  
Aijun An ◽  
Bao Xin Chen ◽  
Junfeng Liu ◽  
...  

1999 ◽  
Vol 7 (1) ◽  
pp. 1-19
Author(s):  
Xiaodong Zhang ◽  
Lin Sun

Shared‐memory and data‐parallel programming models are two important paradigms for scientific applications. Both models provide high‐level program abstractions, and simple and uniform views of network structures. The common features of the two models significantly simplify program coding and debugging for scientific applications. However, the underlining execution and overhead patterns are significantly different between the two models due to their programming constraints, and due to different and complex structures of interconnection networks and systems which support the two models. We performed this experimental study to present implications and comparisons of execution patterns on two commercial architectures. We implemented a standard electromagnetic simulation program (EM) and a linear system solver using the shared‐memory model on the KSR‐1 and the data‐parallel model on the CM‐5. Our objectives are to examine the execution pattern changes required for an implementation transformation between the two models; to study memory access patterns; to address scalability issues; and to investigate relative costs and advantages/disadvantages of using the two models for scientific computations. Our results indicate that the EM program tends to become computation‐intensive in the KSR‐1 shared‐memory system, and memory‐demanding in the CM‐5 data‐parallel system when the systems and the problems are scaled. The EM program, a highly data‐parallel program performed extremely well, and the linear system solver, a highly control‐structured program suffered significantly in the data‐parallel model on the CM‐5. Our study provides further evidence that matching execution patterns of algorithms to parallel architectures would achieve better performance.


2021 ◽  
Author(s):  
Xing Zhao ◽  
Manos Papagelis ◽  
Aijun An ◽  
Bao Xin Chen ◽  
Junfeng Liu ◽  
...  

2001 ◽  
Vol 11 (01) ◽  
pp. 25-40
Author(s):  
CHRISTOPHE CÉRIN ◽  
JEAN-LUC GAUDIOT

We present our experiences in developping and tuning the performance at the user level, of (in core) parallel sorting on homogeneous and non homogeneous clusters with the use of the two available BSP (Bulk Synchronous Parallel model) libraries: BSPLib from Oxford university (UK) and PUB7 from the university of Paderborn (Germany). The paper is mainly about the communication performances of these two libraries and, in more general terms, it compares and summarizes the programming facilities and differences between them.


Sign in / Sign up

Export Citation Format

Share Document