scholarly journals Novel techniques for hardware / software partitioning and emulation

2011 ◽  
Author(s):  
Ιάκωβος Μαυροειδής

Over the last several years, uniprocessor systems, in an effort to overcome the limits of deeperpipelining, instruction-level parallelism and power dissipation, evolved from one processing coreto tens or hundreds of cores. At the same time, multi-chip systems and Systems on Board (SoB),have started giving their place to Systems on Chip (SoC) that exploit the latest nanometertechnologies. This has also caused a tremendous shift in the system development process towardsembedded systems, hardware/software co-design, SoC designs, multi-core designs, and hardwareaccelerators. Nowadays, one of the key issues for continued performance scaling is thedevelopment of advanced CAD tools that can efficiently support the design and verification ofthese new platforms and the requirements of today’s complex applications. This thesis focuses on three important aspects of the system development process: hardware/software partitioning, simulation and verification. Since the time consumed in those tasks is usually a large percentage of the overall development time, speeding them up can significantly reduce the ever important time to market. Hardware emulation on FPGAs has been widely used as a significantly faster and moreaccurate approach for the verification of complex designs than software simulation. In this approach, Hardware Simulation Accelerator and Emulator co-processor units are used to offloadcalculation-intensive tasks from software simulators. One of the biggest problems however is thatthe communication overhead between the software simulator, where the behavioral testbenchusually runs, and the hardware emulator where the Design Under Test (DUT) is emulated, isbecoming a new critical bottleneck. Another problem is that in a hardware emulation environmentit is impossible to bring outside of the chip a large number of internal signals for verificationpurposes. Therefore, on-chip observability has become a significant issue. Finally, one more crucial issue is the decision that has to be made on how to partition the system components into two distinct sets: those that will be implemented in hardware and those that will run in software. Inthis thesis we analyze all the aforementioned problems and propose novel techniques that can beused to attack them. First, we introduce a novel emulation framework that automatically transforms certain HDL parts of the testbench into synthesizable code in order to offload them from the software simulator and, more importantly, minimize the aforementioned communication overhead. In particular, we partition the testbench running on the software simulator into two sections: the testbench HDL code that communicates directly with the DUT and the rest, C-like, testbench code. The former section is transformed into synthesizable code while the latter runs in a general purpose CPU. Next, we extend this architecture by adding multiple fast scan-chain paths in the design in order to provide full circuit observability and controllability on the fly. Finally, we develop a fullyautomated hardware/software partitioning tool that incorporates a novel flow with new costmetrics and functions to provide fast and efficient solutions. The tool employs two separatepartitioning algorithms; Simulated Annealing (SA) and a novel greedy algorithm, the GroupingMapping Partitioning (GMP). Our experiments demonstrate that our methodologies provide cost-effective solutions for the hardware/software partitioning and emulation of large and complex systems.

2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Haoran Wang ◽  
Anton Enders ◽  
John-Alexander Preuss ◽  
Janina Bahnemann ◽  
Alexander Heisterkamp ◽  
...  

Abstract3D printing of microfluidic lab-on-a-chip devices enables rapid prototyping of robust and complex structures. In this work, we designed and fabricated a 3D printed lab-on-a-chip device for fiber-based dual beam optical manipulation. The final 3D printed chip offers three key features, such as (1) an optimized fiber channel design for precise alignment of optical fibers, (2) an optically clear window to visualize the trapping region, and (3) a sample channel which facilitates hydrodynamic focusing of samples. A square zig–zag structure incorporated in the sample channel increases the number of particles at the trapping site and focuses the cells and particles during experiments when operating the chip at low Reynolds number. To evaluate the performance of the device for optical manipulation, we implemented on-chip, fiber-based optical trapping of different-sized microscopic particles and performed trap stiffness measurements. In addition, optical stretching of MCF-7 cells was successfully accomplished for the purpose of studying the effects of a cytochalasin metabolite, pyrichalasin H, on cell elasticity. We observed distinct changes in the deformability of single cells treated with pyrichalasin H compared to untreated cells. These results demonstrate that 3D printed microfluidic lab-on-a-chip devices offer a cost-effective and customizable platform for applications in optical manipulation.


2021 ◽  
Vol 11 (3) ◽  
pp. 1225
Author(s):  
Woohyong Lee ◽  
Jiyoung Lee ◽  
Bo Kyung Park ◽  
R. Young Chul Kim

Geekbench is one of the most referenced cross-platform benchmarks in the mobile world. Most of its workloads are synthetic but some of them aim to simulate real-world behavior. In the mobile world, its microarchitectural behavior has been reported rarely since the hardware profiling features are limited to the public. As a popular mobile performance workload, it is hard to find Geekbench’s microarchitecture characteristics in mobile devices. In this paper, a thorough experimental study of Geekbench performance characterization is reported with detailed performance metrics. This study also identifies mobile system on chip (SoC) microarchitecture impacts, such as the cache subsystem, instruction-level parallelism, and branch performance. After the study, we could understand the bottleneck of workloads, especially in the cache sub-system. This means that the change of data set size directly impacts performance score significantly in some systems and will ruin the fairness of the CPU benchmark. In the experiment, Samsung’s Exynos9820-based platform was used as the tested device with Android Native Development Kit (NDK) built binaries. The Exynos9820 is a superscalar processor capable of dual issuing some instructions. To help performance analysis, we enable the capability to collect performance events with performance monitoring unit (PMU) registers. The PMU is a set of hardware performance counters which are built into microprocessors to store the counts of hardware-related activities. Throughout the experiment, functional and microarchitectural performance profiles were fully studied. This paper describes the details of the mobile performance studies above. In our experiment, the ARM DS5 tool was used for collecting runtime PMU profiles including OS-level performance data. After the comparative study is completed, users will understand more about the mobile architecture behavior, and this will help to evaluate which benchmark is preferable for fair performance comparison.


Author(s):  
Dennis Wolf ◽  
Andreas Engel ◽  
Tajas Ruschke ◽  
Andreas Koch ◽  
Christian Hochberger

AbstractCoarse Grained Reconfigurable Arrays (CGRAs) or Architectures are a concept for hardware accelerators based on the idea of distributing workload over Processing Elements. These processors exploit instruction level parallelism, while being energy efficient due to their simplistic internal structure. However, the incorporation into a complete computing system raises severe challenges at the hardware and software level. This article evaluates a CGRA integrated into a control engineering environment targeting a Xilinx Zynq System on Chip (SoC) in detail. Besides the actual application execution performance, the practicability of the configuration toolchain is validated. Challenges of the real-world integration are discussed and practical insights are highlighted.


2021 ◽  
Vol 5 (4) ◽  
pp. 809-819
Author(s):  
Ahmad Ali Mutezar ◽  
Umniy Salamah

An event is a means for students to improve their soft skill and hard skill. In college, one kind of event that usually held regularly is an exhibition. It is usually held around the universities environtment, but in practice there are still some shortcomings, such as the registration process is done manually, attendance of participants that are not integrated with the system, and unavailability of certificates for participants who have attended the event. Since the outbreak of Covid-19, organizing the events must be done online, so we need a system that can accommodate this. Therefore, this study aims to create an event management system that can manage exhibition event data. Besides, the system is also equipped with a feature to generate an E-Certificate that has a QR Code embedded. The method used in this study is Extreme Programming, with its flexible nature toward changes to facilitate the process of system development. The testing in this study is using black box method, with the test results show that all functional in the system can run well in accordance with user expectations. The use of the Extreme Programming method produces a quality system, because users are involved during the system development process.  


Author(s):  
Michael Leue ◽  
Carlo Luzzi

The San Pedro Bay Ports of Long Beach and Los Angeles continue to provide vital rail connections to the rest of the country. The Rail Enhancement Program sets forth the rail improvements necessary to maintain performance as cargo volumes grow through the year 2035. Implementation of the Rail Enhancement Program has faced hurdles including environmental permitting, funding and competing stakeholder concerns. Cargo growth eased in the years approaching 2010, but the timing of proposed improvements to the rail infrastructure remains critical and challenging. The Rail Enhancement Program is the result of work over the past ten years. Conditions affecting the program have continued to change since the original Rail Master Planning Study of 2000. Updates to the Master Plan have been performed in 2005 and 2010. These documents provide analyses and recommendations for rail improvements to maintain adequate rail service on the Alameda Corridor and through the Port to its rail yards. In developing the Rail Enhancement Program, simulation is used to understand the impacts of increasing cargo volumes on the rail system and to investigate infrastructure and operating improvements required to address deficiencies and to determine improvements to efficiently handle projected traffic. This paper describes the development process with a summary of the analysis methods, resulting proposed rail projects, implementation process and current status of implementation. The steps of the rail system development process include the following: • Evaluation of existing and proposed rail operations; • Conceptual design of over forty potential rail improvement projects; • Analysis of the capacity of existing and proposed facilities; • Scheduling of project development to meet demand; • Estimation of environmental, community and regional impacts and benefits; • Determination of schedule including environmental permit requirements; • Development of project funding plans; and • Preparation of engineering designs and construction documents. The paper will conclude with a summary of the status of key projects from the Rail Enhancement Program. Implementation of the Rail Enhancement Program has included permitting, funding and design efforts on individual projects. The projects currently under development total $1B out of the overall $2B program. The Rail Enhancement Program provides significant benefits to operating efficiencies, environmental impacts and economic impacts. Implementation has been a challenging effort and illustrates the myriad obstacles facing public infrastructure development.


Sign in / Sign up

Export Citation Format

Share Document