Exploitation of parallel processing for implementing high-performance deduction systems

1992 ◽  
Vol 8 (1) ◽  
Author(s):  
Anita Jindal ◽  
Ross Overbeek ◽  
WaldoC. Kabat
2021 ◽  
Vol 17 (11) ◽  
pp. 155014772110331
Author(s):  
Jung-hyun Seo ◽  
HyeongOk Lee

One method to create a high-performance computer is to use parallel processing to connect multiple computers. The structure of the parallel processing system is represented as an interconnection network. Traditionally, the communication links that connect the nodes in the interconnection network use electricity. With the advent of optical communication, however, optical transpose interconnection system networks have emerged, which combine the advantages of electronic communication and optical communication. Optical transpose interconnection system networks use electronic communication for relatively short distances and optical communication for long distances. Regardless of whether the interconnection network uses electronic communication or optical communication, network cost is an important factor among the various measures used for the evaluation of networks. In this article, we first propose a novel optical transpose interconnection system–Petersen-star network with a small network cost and analyze its basic topological properties. Optical transpose interconnection system–Petersen-star network is an undirected graph where the factor graph is Petersen-star network. OTIS–PSN n has the number of nodes 102n, degree n+3, and diameter 6 n − 1. Second, we compare the network cost between optical transpose interconnection system–Petersen-star network and other optical transpose interconnection system networks. Finally, we propose a routing algorithm with a time complexity of 6 n − 1 and a one-to-all broadcasting algorithm with a time complexity of 2 n − 1.


2014 ◽  
Author(s):  
Mehdi Gilaki ◽  
Ilya Avdeev

In this study, we have investigated feasibility of using commercial explicit finite element code LS-DYNA on massively parallel super-computing cluster for accurate modeling of structural impact on battery cells. Physical and numerical lateral impact tests have been conducted on cylindrical cells using a flat rigid drop cart in a custom-built drop test apparatus. The main component of cylindrical cell, jellyroll, is a layered spiral structure which consists of thin layers of electrodes and separator. Two numerical approaches were considered: (1) homogenized model of the cell and (2) heterogeneous (full) 3-D cell model. In the first approach, the jellyroll was considered as a homogeneous material with an effective stress-strain curve obtained through experiments. In the second model, individual layers of anode, cathode and separator were accounted for in the model, leading to extremely complex and computationally expensive finite element model. To overcome limitations of desktop computers, high-performance computing (HPC) techniques on a HPC cluster were needed in order to get the results of transient simulations in a reasonable solution time. We have compared two HPC methods used for this model is shared memory parallel processing (SMP) and massively parallel processing (MPP). Both the homogeneous and the heterogeneous models were considered for parallel simulations utilizing different number of computational nodes and cores and the performance of these models was compared. This work brings us one step closer to accurate modeling of structural impact on the entire battery pack that consists of thousands of cells.


2019 ◽  
pp. 28-31
Author(s):  
E. V. Glivenko ◽  
S. А. Sorokin ◽  
G. N. Petrovа

The article is devoted to the design of high‑performance computing devices for parallel processing of information. The problem of  increasing the productivity of computing facilities by one or several orders of magnitude is considered on the example of the high‑ performance electronic computer M‑10, which was created in the 1970s at the NIIVK. If in a conventional computer, the method  of processing numbers is given by commands, then in M‑10, the methods for processing a function were specified by operators  taken from functional analysis. At the same time, the possibility of parallel processing of an entire information line appeared. Such  systems began to be called «functional operator type machines». The main ideas presented in the article may be of interest to  developers of specialized machines of the new generation, as well as engineers involved in the creation of high‑performance  computing devices using technologies of computing platforms.


Sign in / Sign up

Export Citation Format

Share Document