scholarly journals A Study on Join Operations in MongoDB Preserving Collections Data Models for Future Internet Applications

2019 ◽  
Vol 11 (4) ◽  
pp. 83 ◽  
Author(s):  
Antonio Celesti ◽  
Maria Fazio ◽  
Massimo Villari

Presently, we are observing an explosion of data that need to be stored and processed over the Internet, and characterized by large volume, velocity and variety. For this reason, software developers have begun to look at NoSQL solutions for data storage. However, operations that are trivial in traditional Relational DataBase Management Systems (DBMSs) can become very complex in NoSQL DBMSs. This is the case of the join operation to establish a connection between two or more DB structures, whose construct is not explicitly available in many NoSQL databases. As a consequence, the data model has to be changed or a set of operations have to be performed to address particular queries on data. Thus, open questions are: how do NoSQL solutions work when they have to perform join operations on data that are not natively supported? What is the quality of NoSQL solutions in such cases? In this paper, we deal with such issues specifically considering one of the major NoSQL document oriented DB available on the market: MongoDB. In particular, we discuss an approach to perform join operations at application layer in MongoDB that allows us to preserve data models. We analyse performance of the proposes approach discussing the introduced overhead in comparison with SQL-like DBs.

Author(s):  
Berkay Aydin ◽  
Vijay Akkineni ◽  
Rafal A Angryk

With the ever-growing nature of spatiotemporal data, it is inevitable to use non-relational and distributed database systems for storing massive spatiotemporal datasets. In this chapter, the important aspects of non-relational (NoSQL) databases for storing large-scale spatiotemporal trajectory data are investigated. Mainly, two data storage schemata are proposed for storing trajectories, which are called traditional and partitioned data models. Additionally spatiotemporal and non-spatiotemporal indexing structures are designed for efficiently retrieving data under different usage scenarios. The results of the experiments exhibit the advantages of utilizing data models and indexing structures for various query types.


Author(s):  
Vitor Furlan de Oliveira ◽  
Marcosiris Amorim de Oliveira Pessoa ◽  
Fabrício Junqueira ◽  
Paulo Eigi Miyagi

The data-oriented paradigm has proven to be fundamental for the technological transformation process that characterizes Industry 4.0 (I4.0) so that Big Data & Analytics is considered a technological pillar of this process. The literature reports a series of system architecture proposals that seek to implement the so-called Smart Factory, which is primarily data-driven. Many of these proposals treat data storage solutions as mere entities that support the architecture's functionalities. However, choosing which logical data model to use can significantly affect the performance of the architecture. This work identifies the advantages and disadvantages of relational (SQL) and non-relational (NoSQL) data models for I4.0, taking into account the nature of the data in this process. The characterization of data in the context of I4.0 is based on the five dimensions of Big Data and a standardized format for representing information of assets in the virtual world, the Asset Administration Shell. This work allows identifying appropriate transactional properties and logical data models according to the volume, variety, velocity, veracity, and value of the data. In this way, it is possible to describe the suitability of SQL and NoSQL databases for different scenarios within I4.0.


Machines ◽  
2021 ◽  
Vol 10 (1) ◽  
pp. 20
Author(s):  
Vitor Furlan de Oliveira ◽  
Marcosiris Amorim de Oliveira Pessoa ◽  
Fabrício Junqueira ◽  
Paulo Eigi Miyagi

The data-oriented paradigm has proven to be fundamental for the technological transformation process that characterizes Industry 4.0 (I4.0) so that big data and analytics is considered a technological pillar of this process. The goal of I4.0 is the implementation of the so-called Smart Factory, characterized by Intelligent Manufacturing Systems (IMS) that overcome traditional manufacturing systems in terms of efficiency, flexibility, level of integration, digitalization, and intelligence. The literature reports a series of system architecture proposals for IMS, which are primarily data driven. Many of these proposals treat data storage solutions as mere entities that support the architecture’s functionalities. However, choosing which logical data model to use can significantly affect the performance of the IMS. This work identifies the advantages and disadvantages of relational (SQL) and non-relational (NoSQL) data models for I4.0, considering the nature of the data in this process. The characterization of data in the context of I4.0 is based on the five dimensions of big data and a standardized format for representing information of assets in the virtual world, the Asset Administration Shell. This work allows identifying appropriate transactional properties and logical data models according to the volume, variety, velocity, veracity, and value of the data. In this way, it is possible to describe the suitability of relational and NoSQL databases for different scenarios within I4.0.


2021 ◽  
Vol 18 (2) ◽  
pp. 156-164 ◽  
Author(s):  
Catherine L. Lawson ◽  
Andriy Kryshtafovych ◽  
Paul D. Adams ◽  
Pavel V. Afonine ◽  
Matthew L. Baker ◽  
...  

AbstractThis paper describes outcomes of the 2019 Cryo-EM Model Challenge. The goals were to (1) assess the quality of models that can be produced from cryogenic electron microscopy (cryo-EM) maps using current modeling software, (2) evaluate reproducibility of modeling results from different software developers and users and (3) compare performance of current metrics used for model evaluation, particularly Fit-to-Map metrics, with focus on near-atomic resolution. Our findings demonstrate the relatively high accuracy and reproducibility of cryo-EM models derived by 13 participating teams from four benchmark maps, including three forming a resolution series (1.8 to 3.1 Å). The results permit specific recommendations to be made about validating near-atomic cryo-EM structures both in the context of individual experiments and structure data archives such as the Protein Data Bank. We recommend the adoption of multiple scoring parameters to provide full and objective annotation and assessment of the model, reflective of the observed cryo-EM map density.


Electronics ◽  
2021 ◽  
Vol 10 (10) ◽  
pp. 1139
Author(s):  
Mykola Beshley ◽  
Natalia Kryvinska ◽  
Halyna Beshley ◽  
Oleg Yaremko ◽  
Julia Pyrih

A virtual router model with a static and dynamic resource reconfiguration for future internet networking was developed. This technique allows us to create efficient virtual devices with optimal parameters (queue length, queue overflow management discipline, number of serving devices, mode of serving devices) to ensure the required level of quality of service (QoS). An analytical model of a network device with virtual routers is proposed. By means of the mentioned mathematical representation, it is possible to determine the main parameters of the virtual queue system, which are based on the first in, first out (FIFO) algorithm, in order to analyze the efficiency of network resources utilization, as well as to determine the parameters of QoS flows, for a given intensity of packets arrival at the input interface of the network element. In order to research the guaranteed level of QoS in future telecommunications networks, a simulation model of a packet router with resource virtualization was developed. This model will allow designers to choose the optimal parameters of network equipment for the organization of virtual routers, which, in contrast to the existing principle of service, will provide the necessary quality of service provision to end users in the future network. It is shown that the use of standard static network device virtualization technology is not able to fully provide a guaranteed level of QoS to all present flows in the network by the criterion of minimum delay. An approach for dynamic reconfiguration of network device resources for virtual routers has been proposed, which allows more flexible resource management at certain points in time depending on the input load. Based on the results of the study, it is shown that the dynamic virtualization of the network device provides a guaranteed level of QoS for all transmitted flows. Thus, the obtained results confirm the feasibility of using dynamic reconfiguration of network device resources to improve the quality of service for end users.


2019 ◽  
Vol 35 (S1) ◽  
pp. 85-85
Author(s):  
Sabine Ettinger ◽  
Judit Erdos ◽  
Cecilia De Villiers

IntroductionPatients can provide valuable experience on living with diseases, health-related quality of life, various therapies and relevant outcomes. Their input and perspectives can be helpful in complementing health technology assessment (HTA) processes. The European Network for HTA (EUnetHTA), funded by the European Commission, aims to further advance and standardise patient involvement processes in order to add to the quality and applicability of HTAs and to allow capability building.MethodsDifferent methods for patient involvement in HTAs on non-pharmaceutical technologies were tested: Patient input templates (open questions sent to relevant patient organizations, or published on EUnetHTA website); scoping meeting with patients/patient representatives; one-on-one conversation and group conversation. Applied methods depended on the scope of the HTA and other factors like timelines of HTAs and burden of disease for patients.ResultsPatients were included in eight of sixteen HTAs on non-pharmaceutical technologies. Applied methods were: group conversation (n = 2), scoping meeting (n = 1), patient input templates (n = 4), one-on-one conversation (n = 2,) and other approach (i.e. written feedback on scope n= 2). In some HTAs more than one method was used. Main reasons for not including patients were inability to identify suitable patients or tight timelines. Patients' feedback on health-related quality of life and outcome measures proved most useful in the scoping phase.ConclusionsThe different approaches were useful for complementing HTA processes. Those need to be further tested and evaluated in order to formulate deeper understanding about the impact of patient involvement on HTA. Additionally, feedback from patients that were actively involved in the HTAs should be collected to further improve the involvement methods that should serve as basis for future recommendations post 2020.


2021 ◽  
Vol 11 (1) ◽  
pp. 377
Author(s):  
Michele Scarpiniti ◽  
Enzo Baccarelli ◽  
Alireza Momenzadeh ◽  
Sima Sarv Ahrabi

The recent introduction of the so-called Conditional Neural Networks (CDNNs) with multiple early exits, executed atop virtualized multi-tier Fog platforms, makes feasible the real-time and energy-efficient execution of analytics required by future Internet applications. However, until now, toolkits for the evaluation of energy-vs.-delay performance of the inference phase of CDNNs executed on such platforms, have not been available. Motivated by these considerations, in this contribution, we present DeepFogSim. It is a MATLAB-supported software toolbox aiming at testing the performance of virtualized technological platforms for the real-time distributed execution of the inference phase of CDNNs with early exits under IoT realms. The main peculiar features of the proposed DeepFogSim toolbox are that: (i) it allows the joint dynamic energy-aware optimization of the Fog-hosted computing-networking resources under hard constraints on the tolerated inference delays; (ii) it allows the repeatable and customizable simulation of the resulting energy-delay performance of the overall Fog execution platform; (iii) it allows the dynamic tracking of the performed resource allocation under time-varying operating conditions and/or failure events; and (iv) it is equipped with a user-friendly Graphic User Interface (GUI) that supports a number of graphic formats for data rendering. Some numerical results give evidence for about the actual capabilities of the proposed DeepFogSim toolbox.


2014 ◽  
Vol 513-517 ◽  
pp. 2107-2110 ◽  
Author(s):  
Zhi Jian Diao ◽  
Song Guo

Cloud computing is a novel network-based computing model, in which the cloud infrastructure is constructed in bottom level and provided as the support environment for the applications in upper cloud level. The combination of clouding computing and GIS can improve the performance of GIS, and it can also provide a new prospect of GIS information storage, processing and utilization. By integrating cloud computing and GIS, this paper presented a cloud computing based GIS model based on two features of cloud computing: data storage and transparent custom service. The model contains two layers: service layer and application layer. With this two-layer model, GIS can provide stable and efficient services to end users by optimized network resource allocation of underlying data and services in cloud computing.


Author(s):  
MARIO PIATTINI ◽  
MARCELA GENERO ◽  
LUIS JIMÉNEZ

It is generally accepted in the information system (IS) field that IS quality is highly dependent on the decisions made early in the development life cycle. The construction of conceptual data models is often an important task of this early development. Therefore, improving the quality of conceptual data models will be a major step towards the quality improvement of the IS development. Several quality frameworks for conceptual data models have been proposed, but most of them lack valid quantitative measures in order to evaluate the quality of conceptual data models in an objective way. In this article we will define measures for the structural complexity (internal attribute) of entity relationship diagrams (ERD) and use them for predicting their maintainability (external attribute). We will theoretically validate the proposed metrics following Briand et al.'s framework with the goal of demonstrating the properties that characterise each metric. We will also show how it is possible to predict each of the maintainability sub-characteristics using a prediction model generated using a novel method for induction of fuzzy rules.


2014 ◽  
Vol 11 (4) ◽  
pp. 1271-1289 ◽  
Author(s):  
Maja Pusnik ◽  
Marjan Hericko ◽  
Zoran Budimac ◽  
Bostjan Sumak

In XML Schema development, the quality of XML Schemas is a crucial issue for further steps in the life cycle of an application, closely correlated with the structure of XML Schemas and different building blocks. Current research focuses on measuring complexity of XML Schemas and mainly do not consider other quality aspects. This paper proposes a novel quality measuring approach, based on existing software engineering metrics, additionally defining quality aspect of XML Schemas in the following steps: (1) definition of six schema quality aspects, (2) adoption of 25 directly measurable XML Schema variables, (3) proposition of six composite metrics, applying 25 measured variables and (4) composite metrics validation. An experiment using 250 standard XML Schemas collected from available e-business information systems was conducted. The results illustrate influence of XML Schema characteristics on its quality and evaluate applicability of metrics in the measurement process, a useful tool for software developers while building or adopting XML Schemas.


Sign in / Sign up

Export Citation Format

Share Document