Living on the Edge

2021 ◽  
Vol 2 (3) ◽  
pp. 1-31
Author(s):  
Thilina Buddhika ◽  
Matthew Malensek ◽  
Shrideep Pallickara ◽  
Sangmi Lee Pallickara

Voluminous time-series data streams produced in continuous sensing environments impose challenges pertaining to ingestion, storage, and analytics. In this study, we present a holistic approach based on data sketching to address these issues. We propose a hyper-sketching algorithm that combines discretization and frequency-based sketching to produce compact representations of the multi-feature, time-series data streams. We generate an ensemble of data sketches to make effective use of capabilities at the resource-constrained edge devices, the links over which data are transmitted, and the server pool where this data must be stored. The data sketches can be queried to construct datasets that are amenable to processing using popular analytical engines. We include several performance benchmarks using real-world data from different domains to profile the suitability of our design decisions. The proposed methodology can achieve up to ∼ 13 × and ∼ 2, 207 × reduction in data transfer and energy consumption at edge devices. We observe up to a ∼ 50% improvement in analytical job completion times in addition to the significant improvements in disk and network I/O.

Author(s):  
Kiyohiko Uehara ◽  
◽  
Takumi Koyama ◽  
Kaoru Hirota ◽  

This paper clarifies that inference based on α-cut and generalized mean (α-GEMII) is effective in suppressing consequence deviations. The suppression effect of α-GEMII is numerically evaluated in comparison to conventional inference based on the Compositional Rule of Inference (CRI). CRI-based parallel inference causes discontinuous deviations in the least upper and greatest lower bounds of deduced fuzzy sets even when it models the continuous input-output relation of a system and given facts change continuously. In contrast, α-GEMII can suppress the deviations because of its schemes originally developed for constraint propagation control. In simulations, indices are defined for numerically evaluating the degree to which deduced consequences follow the change in fuzzy outputs of given systems. Simulation results show that α-GEMII is effective in suppressing the deviations, compared to CRI-based parallel inference. In effective use of the schemes for suppressing the consequence deviations, α-GEMII can be applied to nonlinear prediction filters for complex time series, especially with fluctuations that do not always originate from a correlation between time series data.


2016 ◽  
Vol 10 (04) ◽  
pp. 461-501 ◽  
Author(s):  
Om Prasad Patri ◽  
Anand V. Panangadan ◽  
Vikrambhai S. Sorathia ◽  
Viktor K. Prasanna

Detecting and responding to real-world events is an integral part of any enterprise or organization, but Semantic Computing has been largely underutilized for complex event processing (CEP) applications. A primary reason for this gap is the difference in the level of abstraction between the high-level semantic models for events and the low-level raw data values received from sensor data streams. In this work, we investigate the need for Semantic Computing in various aspects of CEP, and intend to bridge this gap by utilizing recent advances in time series analytics and machine learning. We build upon the Process-oriented Event Model, which provides a formal approach to model real-world objects and events, and specifies the process of moving from sensors to events. We extend this model to facilitate Semantic Computing and time series data mining directly over the sensor data, which provides the advantage of automatically learning the required background knowledge without domain expertise. We illustrate the expressive power of our model in case studies from diverse applications, with particular emphasis on non-intrusive load monitoring in smart energy grids. We also demonstrate that this powerful semantic representation is still highly accurate and performs at par with existing approaches for event detection and classification.


2020 ◽  
Vol 4 ◽  
pp. 27
Author(s):  
Daniel M. Weinberger ◽  
Joshua L. Warren

When evaluating the effects of vaccination programs, it is common to estimate changes in rates of disease before and after vaccine introduction. There are a number of related approaches that attempt to adjust for trends unrelated to the vaccine and to detect changes that coincide with introduction. However, characteristics of the data can influence the ability to estimate such a change. These include, but are not limited to, the number of years of available data prior to vaccine introduction, the expected strength of the effect of the intervention, the strength of underlying secular trends, and the amount of unexplained variability in the data. Sources of unexplained variability include model misspecification, epidemics due to unidentified pathogens, and changes in ascertainment or coding practice among others. In this study, we present a simple simulation framework for estimating the power to detect a decline and the precision of these estimates. We use real-world data from a pre-vaccine period to generate simulated time series where the vaccine effect is specified a priori. We present an interactive web-based tool to implement this approach. We also demonstrate the use of this approach using observed data on pneumonia hospitalization from the states in Brazil from a period prior to introduction of pneumococcal vaccines to generate the simulated time series. We relate the power of the hypothesis tests to the number of cases per year and the amount of unexplained variability in the data and demonstrate how fewer years of data influence the results.


Sign in / Sign up

Export Citation Format

Share Document