scholarly journals Cost and Time Economical Planning Algorithm for Scientific Workflows in Cloud Computing

2021 ◽  
Vol 13 (10) ◽  
pp. 263
Author(s):  
Jabanjalin Hilda ◽  
Srimathi Chandrasekaran

A heterogeneous system can be portrayed as a variety of unlike resources that can be locally or geologically spread, which is exploited to implement data-intensive and computationally intensive applications. The competence of implementing the scientific workflow applications on heterogeneous systems is determined by the approaches utilized to allocate the tasks to the proper resources. Cost and time necessity are evolving as different vital concerns of cloud computing environments such as data centers. In the area of scientific workflows, the difficulties of increased cost and time are highly challenging, as they elicit rigorous computational tasks over the communication network. For example, it was discovered that the time to execute a task in an unsuited resource consumes more cost and time in the cloud data centers. In this paper, a new cost- and time-efficient planning algorithm for scientific workflow scheduling has been proposed for heterogeneous systems in the cloud based upon the Predict Optimistic Time and Cost (POTC). The proposed algorithm computes the rank based not only on the completion time of the current task but also on the successor node in the critical path. Under a tight deadline, the running time of the workflow and the transfer cost are reduced by using this technique. The proposed approach is evaluated using true cases of data-exhaustive workflows compared with other algorithms from written works. The test result shows that our proposed method can remarkably decrease the cost and time of the experimented workflows while ensuring a better mapping of the task to the resource. In terms of makespan, speedup, and efficiency, the proposed algorithm surpasses the current existing algorithms—such as Endpoint communication contention-aware List Scheduling Heuristic (ELSH)), Predict Earliest Finish Time (PEFT), Budget-and Deadline-constrained heuristic-based upon HEFT (BDHEFT), Minimal Optimistic Processing Time (MOPT) and Predict Earlier Finish Time (PEFT)—while holding the same time complexity.

2014 ◽  
Vol 1008-1009 ◽  
pp. 1513-1516
Author(s):  
Hai Na Song ◽  
Xiao Qing Zhang ◽  
Zhong Tang He

Cloud computing environment is regarded as a kind of multi-tenant computing mode. With virtulization as a support technology, cloud computing realizes the integration of multiple workloads in one server through the package and seperation of virtual machines. Aiming at the contradiction between the heterogeneous applications and uniform shared resource pool, using the idea of bin packing, the multidimensional resource scheduling problem is analyzed in this paper. We carry out some example analysis in one-dimensional resource scheduling, two-dimensional resource schduling and three-dimensional resource scheduling. The results shows that the resource utilization of cloud data centers will be improved greatly when the resource sheduling is conducted after reorganizing rationally the heterogeneous demands.


Author(s):  
Burak Kantarci ◽  
Hussein T. Mouftah

Cloud computing aims to migrate IT services to distant data centers in order to reduce the dependency of the services on the limited local resources. Cloud computing provides access to distant computing resources via Web services while the end user is not aware of how the IT infrastructure is managed. Besides the novelties and advantages of cloud computing, deployment of a large number of servers and data centers introduces the challenge of high energy consumption. Additionally, transportation of IT services over the Internet backbone accumulates the energy consumption problem of the backbone infrastructure. In this chapter, the authors cover energy-efficient cloud computing studies in the data center involving various aspects such as: reduction of processing, storage, and data center network-related power consumption. They first provide a brief overview of the existing approaches on cool data centers that can be mainly grouped as studies on virtualization techniques, energy-efficient data center network design schemes, and studies that monitor the data center thermal activity by Wireless Sensor Networks (WSNs). The authors also present solutions that aim to reduce energy consumption in data centers by considering the communications aspects over the backbone of large-scale cloud systems.


2019 ◽  
pp. 1360-1369
Author(s):  
Sanjay P. Ahuja ◽  
Karthika Muthiah

Cloud computing is witnessing tremendous growth at one time when climate change and reducing emissions from energy use is gaining attention. With the growth of the cloud, however, comes an increase in demand for energy. There is growing global awareness about reducing greenhouse gas emissions and healthy environments. Green computing in general aims to reduce the consumption of energy and carbon emission and also to recycle and reuse the energy usage in a beneficial and efficient way. Energy consumption is a bottleneck in internet computing technology. Green cloud computing related technology arose as an improvement to cloud computing. Cloud data centers consume inordinate amounts of energy and have significant CO2 emissions as they have a huge network of servers. Furthermore, these data centers are tightly linked to provide high performance services, outsourcing and sharing resources to multiple users through the internet. This paper gives an overview about green cloud computing and its evolution, surveys related work, discusses associated integrated green cloud architecture – Green Cloud Framework, innovations, and technologies, and highlights future work and challenges that need to be addressed to sustain an eco-friendly cloud computing environment that is poised for significant growth.


2019 ◽  
Vol 9 (17) ◽  
pp. 3550 ◽  
Author(s):  
A-Young Son ◽  
Eui-Nam Huh

With the rapid increase in the development of the cloud data centers, it is expected that massive data will be generated, which will decrease service response time for the cloud data centers. To improve the service response time, distributed cloud computing has been designed and researched for placement and migration from mobile devices close to edge servers that have secure resource computing. However, most of the related studies did not provide sufficient service efficiency for multi-objective factors such as energy efficiency, resource efficiency, and performance improvement. In addition, most of the existing approaches did not consider various metrics. Thus, to maximize energy efficiency, maximize performance, and reduce costs, we consider multi-metric factors by combining decision methods, according to user requirements. In order to satisfy the user’s requirements based on service, we propose an efficient service placement system named fuzzy- analytical hierarchical process and then analyze the metric that enables the decision and selection of a machine in the distributed cloud environment. Lastly, using different placement schemes, we demonstrate the performance of the proposed scheme.


2019 ◽  
Vol 16 (9) ◽  
pp. 3989-3994
Author(s):  
Jaspreet Singh ◽  
Deepali Gupta ◽  
Neha Sharma

Nowadays, Cloud computing is developing quickly and customers are requesting more administrations and superior outcomes. In the cloud domain, load balancing has turned into an extremely intriguing and crucial research area. Numbers of algorithms were recommended to give proficient mechanism for distributing the cloud user’s requests for accessing pool cloud resources. Also load balancing in cloud should provide notable functional benefits to cloud users and at the same time should prove out to be eminent for cloud services providers. In this paper, the pre-existing load balancing techniques are explored. The paper intends to provide landscape for classification of distinct load balancing algorithms based upon the several parameters and also address performance assessment bound to various load balancing algorithms. The comparative assessment of various load balancing algorithms will helps in proposing a competent load balancing technique for intensify the performance of cloud data centers.


Author(s):  
Deepika T. ◽  
Prakash P.

The flourishing development of the cloud computing paradigm provides several services in the industrial business world. Power consumption by cloud data centers is one of the crucial issues for service providers in the domain of cloud computing. Pursuant to the rapid technology enhancements in cloud environments and data centers augmentations, power utilization in data centers is expected to grow unabated. A diverse set of numerous connected devices, engaged with the ubiquitous cloud, results in unprecedented power utilization by the data centers, accompanied by increased carbon footprints. Nearly a million physical machines (PM) are running all over the data centers, along with (5 – 6) million virtual machines (VM). In the next five years, the power needs of this domain are expected to spiral up to 5% of global power production. The virtual machine power consumption reduction impacts the diminishing of the PM’s power, however further changing in power consumption of data center year by year, to aid the cloud vendors using prediction methods. The sudden fluctuation in power utilization will cause power outage in the cloud data centers. This paper aims to forecast the VM power consumption with the help of regressive predictive analysis, one of the Machine Learning (ML) techniques. The potency of this approach to make better predictions of future value, using Multi-layer Perceptron (MLP) regressor which provides 91% of accuracy during the prediction process.


Sensors ◽  
2021 ◽  
Vol 21 (21) ◽  
pp. 7238
Author(s):  
Zulfiqar Ahmad ◽  
Ali Imran Jehangiri ◽  
Mohammed Alaa Ala’anzy ◽  
Mohamed Othman ◽  
Arif Iqbal Umar

Cloud computing is a fully fledged, matured and flexible computing paradigm that provides services to scientific and business applications in a subscription-based environment. Scientific applications such as Montage and CyberShake are organized scientific workflows with data and compute-intensive tasks and also have some special characteristics. These characteristics include the tasks of scientific workflows that are executed in terms of integration, disintegration, pipeline, and parallelism, and thus require special attention to task management and data-oriented resource scheduling and management. The tasks executed during pipeline are considered as bottleneck executions, the failure of which result in the wholly futile execution, which requires a fault-tolerant-aware execution. The tasks executed during parallelism require similar instances of cloud resources, and thus, cluster-based execution may upgrade the system performance in terms of make-span and execution cost. Therefore, this research work presents a cluster-based, fault-tolerant and data-intensive (CFD) scheduling for scientific applications in cloud environments. The CFD strategy addresses the data intensiveness of tasks of scientific workflows with cluster-based, fault-tolerant mechanisms. The Montage scientific workflow is considered as a simulation and the results of the CFD strategy were compared with three well-known heuristic scheduling policies: (a) MCT, (b) Max-min, and (c) Min-min. The simulation results showed that the CFD strategy reduced the make-span by 14.28%, 20.37%, and 11.77%, respectively, as compared with the existing three policies. Similarly, the CFD reduces the execution cost by 1.27%, 5.3%, and 2.21%, respectively, as compared with the existing three policies. In case of the CFD strategy, the SLA is not violated with regard to time and cost constraints, whereas it is violated by the existing policies numerous times.


Author(s):  
Raghi K.R K R

Cloud computing data centers are growing rapidly in both number and capacity to meet the increasing demands for highly-responsive computing and massive storage. Such data centers consume enormous amounts of electrical energy resulting in high operating costs and carbon dioxide emissions. The reason for this extremely high energy consumption is not just the quantity of computing resources and the power inefficiency of hardware, but rather lies in the inefficient usage of these resources. Virtual Machine [VM] consolidation involves live migration of VMs hence the capability of transferring a VM between physical servers with a close to zero down time. It is an effective way to improve the utilization of resources and increase energy efficiency in cloud data centers. VM consolidation consists of host overload/under load detection, VM selection and VM placement. In Our Proposed Model We are going to use Roulette-Wheel Selection Strategy, Where the VM selects the Instance type and Physical Machine [PM] using Roulette-Wheel Selection Mechanism Keywords—searchable encryption, dynamic update, cloud computing


Author(s):  
Jasraj Meena ◽  
Manu Vardhan

Cloud computing is used to deliver IT resources over the internet. Due to the popularity of cloud computing, nowadays, most of the scientific workflows are shifted towards this environment. There are lots of algorithms has been proposed in the literature to schedule scientific workflows in the cloud, but their execution cost is very high as well as they are not meeting the user-defined deadline constraint. This paper focuses on satisfying the userdefined deadline of a scientific workflow while minimizing the total execution cost. So, to achieve this, we have proposed a Cost-Effective under Deadline (CEuD) constraint workflow scheduling algorithm. The proposed CEuD algorithm considers all the essential features of Cloud and resolves the major issues such as performance variation, and acquisition delay. We have compared the proposed CEuD algorithm with the existing literature algorithms for scientific workflows (i.e., Montage, Epigenomics, and CyberShake) and getting better results for minimizing the overall execution cost of the workflow while satisfying the user-defined deadline.


Author(s):  
Atefeh Khosravi ◽  
Rajkumar Buyya

Cloud computing provides on-demand access to computing resources for users across the world. It offers services on a pay-as-you-go model through data center sites that are scattered across diverse geographies. However, cloud data centers consume huge amount of electricity and leave high amount of carbon footprint in the ecosystem. This makes data centers responsible for 2% of the global CO2 emission. Therefore, having energy and carbon-efficient techniques for resource management in distributed cloud data centers is inevitable. This chapter presents a taxonomy and classifies the existing research works based on their target system, objective, and the technique they use for resource management in achieving a green cloud computing environment. Finally, it discusses how each work addresses the issue of energy and carbon-efficiency and also provides an insight into future directions.


Sign in / Sign up

Export Citation Format

Share Document