scholarly journals Using Multicore Technologies to Speed Up Complex Simulations of Population Evolution

2013 ◽  
Vol 2013 ◽  
pp. 1-9
Author(s):  
Mauricio Guevara-Souza ◽  
Edgar E. Vallejo

We explore with the use of multicore processing technologies for conducting simulations on population replacement of disease vectors. In our model, a native population of simulated vectors is inoculated with a small exogenous population of vectors that have been infected with theWolbachiabacteria, which confers immunity to the disease. We conducted a series of computational simulations to study the conditions required by the invading population to take over the native population. Given the computational burden of this study, we decided to take advantage of modern multicore processor technologies for reducing the time required for the simulations. Overall, the results seem promising both in terms of the application and the use of multicore technologies.

Author(s):  
Nikolaos Athanasios Anagnostopoulos ◽  
Tolga Arul ◽  
Yufan Fan ◽  
Christian Hatzfeld ◽  
André Schaller ◽  
...  

Physical Unclonable Functions (PUFs) based on the retention times of the cells of a Dynamic Random Access Memory (DRAM) can be utilised for the implementation of cost-efficient and lightweight cryptographic protocols. However, as recent work has demonstrated, the times needed in order to generate their responses may prohibit their widespread usage. In order to address this issue, the Row Hammer PUF has been proposed by Schaller et al. [1], which leverages the row hammer effect in DRAM modules to reduce the retention times of their cells and, therefore, significantly speed up the generation times for the responses of PUFs based on these retention times. In this work, we extend the work of Schaller et al. by presenting a run-time accessible implementation of this PUF and further reducing the time required for the generation of its responses. Additionally, we also provide a more thorough investigation of the effects of temperature variations on the the Row Hammer PUF and briefly discuss potential statistical relationships between the cells used to implement it. As our results prove, the Row Hammer PUF could potentially provide an adequate level of security for Commercial Off-The-Shelf (COTS) devices, if its dependency on temperature is mitigated, and, may therefore, be commercially adopted in the near future.


1969 ◽  
Vol 39 (6) ◽  
pp. 497-504 ◽  
Author(s):  
Norman R. S. Hollies ◽  
Steven R. Chafitz ◽  
Karen A. Farquhar

The impregnation of cotton fabrics with a solution consisting of strong acid and a combination of N-methylol resins having polymer-forming and cross-linking properties distinguishes the wet-fixation system from conventional durable-press processes, and this finish results in an improved balance of smoothness and strength properties during wear and laundering. Swelling of the fibers in a steam atmosphere, following padding in the resin solution can serve to speed up the impregnation process. In addition, with controls to minimize resin migration back to the fiber surface, steaming can substantially improve the efficiency of the use of resin for producing these smooth drying properties. The degree of penetration of resin is influenced by a number of process variables, such as predrying before steaming, steaming time, fabric tension and rapidity of neutralization. The optimum in fabric performance is achieved with both sufficient resin of both types in the fiber system and even distribution of resin within the individual fibers, Steaming acts to improve both these factors over that achieved in conventional hot wet fixation and so reduces the time required for wet fixation by a factor of 20–30. There is a corresponding increase in efficiency of resin use so that resin solids in the bath can be reduced two- to three-fold. These findings appear to have general application to a variety of cotton finishing processes involving fiber impregnation with reactive resins.


2020 ◽  
Vol 1 (3) ◽  
pp. 1056-1067 ◽  
Author(s):  
Serge P. J. M. Horbach

In times of public crises, including the current COVID-19 pandemic, rapid dissemination of relevant scientific knowledge is of paramount importance. The duration of scholarly journals’ publication process is one of the main factors that may hinder quick delivery of new information. Following initiatives of medical journals to accelerate their publication process, this study assesses whether medical journals have managed to speed up their publication process for coronavirus-related articles. It studies the duration of 14 medical journals’ publication processes both during and prior to the current pandemic. Assessing 669 articles, the study concludes that medical journals have indeed strongly accelerated their publication process for coronavirus-related articles since the outbreak of the pandemic: The time between submission and publication has decreased on average by 49%. The largest decrease in number of days between submission and publication of articles was due to a decrease in time required for peer review. For articles not related to COVID-19, no acceleration of the publication process is found. While the acceleration of the publication process is laudable from the perspective of quick information dissemination, it also may raise concerns relating to the quality of the peer review process and of the resulting publications.


VLSI Design ◽  
2012 ◽  
Vol 2012 ◽  
pp. 1-11
Author(s):  
M. Walton ◽  
O. Ahmed ◽  
G. Grewal ◽  
S. Areibi

Scatter Search is an effective and established population-based metaheuristic that has been used to solve a variety of hard optimization problems. However, the time required to find high-quality solutions can become prohibitive as problem sizes grow. In this paper, we present a hardware implementation of Scatter Search on a field-programmable gate array (FPGA). Our objective is to improve the run time of Scatter Search by exploiting the potentially massive performance benefits that are available through the native parallelism in hardware. When implementing Scatter Search we employ two different high-level languages (HLLs): Handel-C and Impulse-C. Our empirical results show that by effectively exploiting source-code optimizations, data parallelism, and pipelining, a 28x speed up over software can be achieved.


2018 ◽  
Vol 232 ◽  
pp. 02028
Author(s):  
Bowei Liu ◽  
Yi Sun ◽  
Mei Wang

The wind loads could be increased by the complex hilly terrain. In order to consider this threat, computational simulations about wind pressures in buildings in hilly terrain and surrounding buildings were finished. In the CFD model, surrounding buildings within 500 meters and hilly terrain within 700 meters were established, and total 16 cases including different sensitive wind directions were simulated. The results showed that, because of the complex situation of surrounding hilly terrain and interfered buildings, wind loads in the building group under different direction show opposite characteristics. The speed-up effects by the hilly terrain increased the windward mean pressures when the flow comes from the north, while the interference effects by the surrounding buildings decreased them when the flow comes from the south.


2021 ◽  
Vol 11 (18) ◽  
pp. 8750
Author(s):  
Styliani Verykokou ◽  
Argyro-Maria Boutsi ◽  
Charalabos Ioannidis

Mobile Augmented Reality (MAR) is designed to keep pace with high-end mobile computing and their powerful sensors. This evolution excludes users with low-end devices and network constraints. This article presents ModAR, a hybrid Android prototype that expands the MAR experience to the aforementioned target group. It combines feature-based image matching and pose estimation with fast rendering of 3D textured models. Planar objects of the real environment are used as pattern images for overlaying users’ meshes or the app’s default ones. Since ModAR is based on the OpenCV C++ library at Android NDK and OpenGL ES 2.0 graphics API, there are no dependencies on additional software, operating system version or model-specific hardware. The developed 3D graphics engine implements optimized vertex-data rendering with a combination of data grouping, synchronization, sub-texture compression and instancing for limited CPU/GPU resources and a single-threaded approach. It achieves up to 3 × speed-up compared to standard index rendering, and AR overlay of a 50 K vertices 3D model in less than 30 s. Several deployment scenarios on pose estimation demonstrate that the oriented FAST detector with an upper threshold of features per frame combined with the ORB descriptor yield best results in terms of robustness and efficiency, achieving a 90% reduction of image matching time compared to the time required by the AGAST detector and the BRISK descriptor, corresponding to pattern recognition accuracy of above 90% for a wide range of scale changes, regardless of any in-plane rotations and partial occlusions of the pattern.


Author(s):  
Ying Sheng ◽  
Yoni Zohar ◽  
Christophe Ringeissen ◽  
Andrew Reynolds ◽  
Clark Barrett ◽  
...  

AbstractWe make two contributions to the study of polite combination in satisfiability modulo theories. The first is a separation between politeness and strong politeness, by presenting a polite theory that is not strongly polite. This result shows that proving strong politeness (which is often harder than proving politeness) is sometimes needed in order to use polite combination. The second contribution is an optimization to the polite combination method, obtained by borrowing from the Nelson-Oppen method. The Nelson-Oppen method is based on guessing arrangements over shared variables. In contrast, polite combination requires an arrangement over all variables of the shared sorts. We show that when using polite combination, if the other theory is stably infinite with respect to a shared sort, only the shared variables of that sort need be considered in arrangements, as in the Nelson-Oppen method. The time required to reason about arrangements is exponential in the worst case, so reducing the number of variables considered has the potential to improve performance significantly. We show preliminary evidence for this by demonstrating a speed-up on a smart contract verification benchmark.


2019 ◽  
pp. 75-81
Author(s):  
Kailash Kalare ◽  
Jitendra Tembhurne

The sequence alignment comprises to identify similarities and dissimilarities between two given sequences. In this paper, we propose a work distribution strategy for the implementation of DNA global sequence alignment. The main objective of this work is to minimize the execution time required for DNA global alignment of large biological sequences. The proposed approach dealt the issues with the memory optimizations and minimization of execution time. We considered the biological sequences of different size to fit into the global memory of the system. The proposed strategy is implemented in shared memory architecture using OpenMP programming for large biological sequences. Parallelization using OpenMP directive is relatively easy and execute the code fast. We experimented on the Dell Precision Tower 7910 with Intel Xeon processor with 32GB RAM and 28 CPU cores. The efficient use of global memory and cache memory optimization dominate the results of execution time. The results demonstrate the significantly high speed up using OpenMP as compared with other implementations.


2021 ◽  
Vol 2021 (6) ◽  
pp. 5381-5390
Author(s):  
MILAN MIHOLA ◽  
◽  
ZDENEK ZEMAN ◽  
DAVID FOJTIK ◽  
◽  
...  

The design of mechatronic devices is a demanding process not only in terms of the time required but also of the demands placed on the knowledge and experience of development workers. The aim of this research and development was to create suitable procedures, algorithms, and databases of 3D models, with the help of which could this process be significantly shortened and simplified. The results of the development are a software tool for the design of electric drive units, procedures for creating 3D models with the possibility of using the SolidWorks software API, methods for automating the creation of assemblies of 3D models and a description of the knowledge database in which various data and algorithms are stored. The benefits of the proposed procedures, the Drive Picker software tool, and the knowledge database, are demonstrated on the design of a robotic arm with 5 degrees of freedom. Despite the complexity of the preparation of documents needed, it turns out that the chosen approach can significantly speed up and simplify the design of mechatronic devices.


2020 ◽  
Vol 245 ◽  
pp. 02001
Author(s):  
Marilena Bandieramonte ◽  
John Derek Chapman ◽  
Justin Chiu ◽  
Heather Gray ◽  
Miha Muskinja

Estimations of the CPU resources that will be needed to produce simulated data for the future runs of the ATLAS experiment at the LHC, indicate a compelling need to speed-up the process to reduce the computational time required. While different fast simulation projects are ongoing, full Geant4 based simulation will still be heavily used and is expected to consume the biggest portion of the total estimated processing time. In order to run effectively on modern architectures and profit from multi-core designs a migration of the Athena framework to a multi-threading processing model was performed. A multi-threaded simulation based on AthenaMT and Geant4MT, enables substantial decreases in the memory footprint of jobs, largely from shared geometry and cross-section tables. This approach scales better with respect to the multi-processing approach (AthenaMP) especially on the architectures that are foreseen to be used in the next LHC runs. In these proceedings we report about the status of the multi-threaded simulation in ATLAS, focusing on the different challenges of its validation process. We demonstrate the different tools and strategies that have been used for debugging multi-threaded runs versus the corresponding sequential ones, in order to have a fully reproducible and consistent simulation result.


Sign in / Sign up

Export Citation Format

Share Document