scholarly journals Modeling and Analyzing Offloading Strategies of IoT Applications over Edge Computing and Joint Clouds

Symmetry ◽  
2021 ◽  
Vol 13 (3) ◽  
pp. 402
Author(s):  
Jaber Almutairi ◽  
Mohammad Aldossary

Internet of Things (IoT) is swiftly evolving into a disruptive technology in recent years. For enhancing customer experience and accelerating job execution, IoT task offloading enables mobile end devices to release heavy computation and storage to the resource-rich nodes in collaborative Edges or Clouds. However, how different service architecture and offloading strategies quantitatively impact the end-to-end performance of IoT applications is still far from known particularly given a dynamic and unpredictable assortment of interconnected virtual and physical devices. This paper exploits potential network performance that manifests within the edge-cloud environment, then investigates and compares the impacts of two types of architectures: Loosely-Coupled (LC) and Orchestrator-Enabled (OE). Further, it introduces three customized offloading strategies in order to handle various requirements for IoT latency-sensitive applications. Through comparative experiments, we observed that the computational requirements exerts more influence on the IoT application’s performance compared to the communication requirement. However, when the system scales up to accommodate more IoT devices, communication bandwidth will turn to be the dominant resource and becomes the essential factor that will directly impact the overall performance. Thus, orchestration is a necessary procedure to encompass optimized solutions under different constraints for optimal offloading placement.

2020 ◽  
Vol 2020 ◽  
pp. 1-11
Author(s):  
Wanli Zhang ◽  
Xianwei Li ◽  
Liang Zhao ◽  
Xiaoying Yang

Network performance is of great importance for processing Internet of Things (IoT) applications in the fifth-generation (5G) communication system. With the increasing number of the devices, how network services should be provided with better performances is becoming a pressing issue. The static resource allocation of wireless networks is becoming a bottleneck for the emerging IoT applications. As a potential solution, network virtualization is considered a promising approach to enhancing the network performance and solving the bottleneck issue. In this paper, the problem of wireless network virtualization is investigated where one wireless infrastructure provider (WIP), mobile virtual network operators (MVNOs), and IoT devices coexist. In the system model under consideration, with the help of a software-defined network (SDN) controller, the WIP can divide and reconfigure its radio frequency bands to radio frequency slices. Then, two MVNOs, MVNO1 and MVNO2, can lease these frequency slices from the WIP and then provide IoT network services to IoT users under competition. We apply a two-stage Stackelberg game to investigate and analyze the relationship between the two MVNOs and IoT users, where MVNO1 and MVNO2 firstly try to maximize their profits by setting the optimal network service prices. Then, IoT users make decisions on which network service they should select according to the performances and prices of network services. Two competition cases between MVNO1 and MVNO2 are considered, namely, Stackelberg game (SG) where MVNO1 is the leader whose price of network service is set firstly and MVNO2 is the follower whose network service price is set later and noncooperative strategic game (NSG) under which the service prices of MVNO1 and MVNO2 are simultaneously set. Each IoT user decides whether and which MVNO to select on the basis of the network service prices and qualities. The numerical results are provided to show the effectiveness of our game model and the proposed solution method.


Symmetry ◽  
2019 ◽  
Vol 12 (1) ◽  
pp. 3
Author(s):  
Van-Nam Pham ◽  
VanDung Nguyen ◽  
Tri D. T. Nguyen ◽  
Eui-Nam Huh

Computing services for the Internet-of-Things (IoT) play a vital role for widespread IoT deployment. A hierarchy of Edge-Cloud publish/subscribe (pub/sub) broker overlay networks that support latency-sensitive IoT applications in a scalable manner is introduced. In addition, we design algorithms to cluster edge pub/sub brokers based on topic similarities and geolocations to enhance data dissemination among end-to-end IoT devices. The proposed model is designed to provide low delay data dissemination and effectively save network traffic among brokers. In the proposed model, IoT devices running pub/sub client applications periodically send collected data, organized as a hierarchy of topics, to their closest edge pub/sub brokers. Then, the data are processed/analyzed at edge nodes to make controlling decisions promptly replying to the IoT devices and/or aggregated for further delivery to other interested edge brokers or to cloud brokers for long-term processing, analysis, and storage. Extensive simulation results demonstrate that our proposal achieves the best data delivery latency compared to two baseline schemes, a classical Cloud-based pub/sub scheme and an Edge-Cloud pub/sub scheme. Considering the similar Edge-Cloud technique, the proposed scheme outperforms PubSubCoord-alike in terms of relay traffic ratio among brokers. Therefore, our proposal can adapt well to support wide-scale latency-sensitive IoT applications.


Electronics ◽  
2020 ◽  
Vol 9 (6) ◽  
pp. 1050 ◽  
Author(s):  
Yustus Eko Oktian ◽  
Sang-Gon Lee ◽  
Hoon Jae Lee

Many researchers challenge the possibility of using blockchain and smart contracts to disrupt the Internet of Things (IoT) architecture because of their security and decentralization guarantees. However, the state-of-the-art blockchain architecture is not scalable enough to satisfy the requirements of massive data traffics in the IoT environment. The main reason for this issue is one needs to choose the consensus trade-off between either coping with a high throughput or a high number of nodes. Consequently, this issue prevents the applicability of blockchain for IoT use cases. In this paper, we propose a scalable two-tiered hierarchical blockchain architecture for IoT. The first tier is a Core Engine, which is based on a Practical Byzantine Fault Tolerance (PBFT) consensus to cope with a high throughput, that supervises the underlying subordinate engines (sub-engines) as its second tier. This second tier comprises of the Payment, Compute, and Storage Engine, respectively. We can deploy multiple instances of these sub-engines as many as we need and as local as possible near to the IoT domains, where IoT devices reside, to cope with a high number of nodes. Furthermore, to further extend the scalability of the proposed architecture, we also provide additional scalability features on the Core Engine such as request aggregation, request prioritization, as well as sub-engine parallelism. We implement all of our engines and expose them to IoT applications through the Engine APIs. With these APIs, developers can build and run IoT applications in our architecture. Our evaluation results show that our proposed features on the Core Engine can indeed enhance the overall performance of our architecture. Moreover, based on our proof-of-concept IoT car rental application, we also show that the interoperability between sub-engines through the Core Engine is possible, even when the particular sub-engine is under sub-engine parallelism.


Sensors ◽  
2020 ◽  
Vol 20 (22) ◽  
pp. 6441 ◽  
Author(s):  
Salam Hamdan ◽  
Moussa Ayyash ◽  
Sufyan Almajali

The rapid growth of the Internet of Things (IoT) applications and their interference with our daily life tasks have led to a large number of IoT devices and enormous sizes of IoT-generated data. The resources of IoT devices are limited; therefore, the processing and storing IoT data in these devices are inefficient. Traditional cloud-computing resources are used to partially handle some of the IoT resource-limitation issues; however, using the resources in cloud centers leads to other issues, such as latency in time-critical IoT applications. Therefore, edge-cloud-computing technology has recently evolved. This technology allows for data processing and storage at the edge of the network. This paper studies, in-depth, edge-computing architectures for IoT (ECAs-IoT), and then classifies them according to different factors such as data placement, orchestration services, security, and big data. Besides, the paper studies each architecture in depth and compares them according to various features. Additionally, ECAs-IoT is mapped according to two existing IoT layered models, which helps in identifying the capabilities, features, and gaps of every architecture. Moreover, the paper presents the most important limitations of existing ECAs-IoT and recommends solutions to them. Furthermore, this survey details the IoT applications in the edge-computing domain. Lastly, the paper recommends four different scenarios for using ECAs-IoT by IoT applications.


Author(s):  
MAHESH KALUTI

Despite the technical changes and enormous day by day upgradiation in the field of mobile computing the smart devices as well as IOT devices had experienced tremendous technical glitch, which narrow’s the life span and survivability of small scale processing devices. Today, end users are becoming more demanding and are expecting to run computational intensive tasks on their Smart phone devices and IOT devices. Therefore, virtual cloud computing (VCC) integrates local device computing and Cloud Computing (CC) in order to extend computational capabilities of smart phone devices and IOT devices using cloud offloading techniques. Computation Offloading tackles limitations of Smart phone devices and IOT devices such as limited battery duration, limited computational capabilities, and limited storage capacity by offloading the execution and workload to cloud which has better systems with better computation and storage capabilities. This paper aims to present the techniques to offload computational intensive tasks to cloud framework and analyses them along with traditional local execution techniques and their issues. Furthermore, it explores other important parameters based on which the applications are implemented such as offloading technique and partitioning of tasks.


2021 ◽  
Vol 10 (1) ◽  
pp. 13
Author(s):  
Claudia Campolo ◽  
Giacomo Genovese ◽  
Antonio Iera ◽  
Antonella Molinaro

Several Internet of Things (IoT) applications are booming which rely on advanced artificial intelligence (AI) and, in particular, machine learning (ML) algorithms to assist the users and make decisions on their behalf in a large variety of contexts, such as smart homes, smart cities, smart factories. Although the traditional approach is to deploy such compute-intensive algorithms into the centralized cloud, the recent proliferation of low-cost, AI-powered microcontrollers and consumer devices paves the way for having the intelligence pervasively spread along the cloud-to-things continuum. The take off of such a promising vision may be hurdled by the resource constraints of IoT devices and by the heterogeneity of (mostly proprietary) AI-embedded software and hardware platforms. In this paper, we propose a solution for the AI distributed deployment at the deep edge, which lays its foundation in the IoT virtualization concept. We design a virtualization layer hosted at the network edge that is in charge of the semantic description of AI-embedded IoT devices, and, hence, it can expose as well as augment their cognitive capabilities in order to feed intelligent IoT applications. The proposal has been mainly devised with the twofold aim of (i) relieving the pressure on constrained devices that are solicited by multiple parties interested in accessing their generated data and inference, and (ii) and targeting interoperability among AI-powered platforms. A Proof-of-Concept (PoC) is provided to showcase the viability and advantages of the proposed solution.


Author(s):  
Jaber Almutairi ◽  
Mohammad Aldossary

AbstractRecently, the number of Internet of Things (IoT) devices connected to the Internet has increased dramatically as well as the data produced by these devices. This would require offloading IoT tasks to release heavy computation and storage to the resource-rich nodes such as Edge Computing and Cloud Computing. Although Edge Computing is a promising enabler for latency-sensitive related issues, its deployment produces new challenges. Besides, different service architectures and offloading strategies have a different impact on the service time performance of IoT applications. Therefore, this paper presents a novel approach for task offloading in an Edge-Cloud system in order to minimize the overall service time for latency-sensitive applications. This approach adopts fuzzy logic algorithms, considering application characteristics (e.g., CPU demand, network demand and delay sensitivity) as well as resource utilization and resource heterogeneity. A number of simulation experiments are conducted to evaluate the proposed approach with other related approaches, where it was found to improve the overall service time for latency-sensitive applications and utilize the edge-cloud resources effectively. Also, the results show that different offloading decisions within the Edge-Cloud system can lead to various service time due to the computational resources and communications types.


Sensors ◽  
2021 ◽  
Vol 21 (8) ◽  
pp. 2681
Author(s):  
Kedir Mamo Besher ◽  
Juan Ivan Nieto-Hipolito ◽  
Raymundo Buenrostro-Mariscal ◽  
Mohammed Zamshed Ali

With constantly increasing demand in connected society Internet of Things (IoT) network is frequently becoming congested. IoT sensor devices lose more power while transmitting data through congested IoT networks. Currently, in most scenarios, the distributed IoT devices in use have no effective spectrum based power management, and have no guarantee of a long term battery life while transmitting data through congested IoT networks. This puts user information at risk, which could lead to loss of important information in communication. In this paper, we studied the extra power consumed due to retransmission of IoT data packet and bad communication channel management in a congested IoT network. We propose a spectrum based power management solution that scans channel conditions when needed and utilizes the lowest congested channel for IoT packet routing. It also effectively measured power consumed in idle, connected, paging and synchronization status of a standard IoT device in a congested IoT network. In our proposed solution, a Freescale Freedom Development Board (FREDEVPLA) is used for managing channel related parameters. While supervising the congestion level and coordinating channel allocation at the FREDEVPLA level, our system configures MAC and Physical layer of IoT devices such that it provides the outstanding power utilization based on the operating network in connected mode compared to the basic IoT standard. A model has been set up and tested using freescale launchpads. Test data show that battery life of IoT devices using proposed spectrum based power management increases by at least 30% more than non-spectrum based power management methods embedded within IoT devices itself. Finally, we compared our results with the basic IoT standard, IEEE802.15.4. Furthermore, the proposed system saves lot of memory for IoT devices, improves overall IoT network performance, and above all, decrease the risk of losing data packets in communication. The detail analysis in this paper also opens up multiple avenues for further research in future use of channel scanning by FREDEVPLA board.


Sensors ◽  
2019 ◽  
Vol 19 (6) ◽  
pp. 1339 ◽  
Author(s):  
Hasan Islam ◽  
Dmitrij Lagutin ◽  
Antti Ylä-Jääski ◽  
Nikos Fotiou ◽  
Andrei Gurtov

The Constrained Application Protocol (CoAP) is a specialized web transfer protocol which is intended to be used for constrained networks and devices. CoAP and its extensions (e.g., CoAP observe and group communication) provide the potential for developing novel applications in the Internet-of-Things (IoT). However, a full-fledged CoAP-based application may require significant computing capability, power, and storage capacity in IoT devices. To address these challenges, we present the design, implementation, and experimentation with the CoAP handler which provides transparent CoAP services through the ICN core network. In addition, we demonstrate how the CoAP traffic over an ICN network can unleash the full potential of the CoAP, shifting both overhead and complexity from the (constrained) endpoints to the ICN network. The experiments prove that the CoAP Handler helps to decrease the required computation complexity, communication overhead, and state management of the CoAP server.


Sign in / Sign up

Export Citation Format

Share Document