aggregation queries
Recently Published Documents


TOTAL DOCUMENTS

60
(FIVE YEARS 4)

H-INDEX

10
(FIVE YEARS 0)

2021 ◽  
Vol 14 (11) ◽  
pp. 2341-2354
Author(s):  
Daniel Kang ◽  
John Guibas ◽  
Peter Bailis ◽  
Tatsunori Hashimoto ◽  
Yi Sun ◽  
...  

Researchers and industry analysts are increasingly interested in computing aggregation queries over large, unstructured datasets with selective predicates that are computed using expensive deep neural networks (DNNs). As these DNNs are expensive and because many applications can tolerate approximate answers, analysts are interested in accelerating these queries via approximations. Unfortunately, standard approximate query processing techniques to accelerate such queries are not applicable because they assume the result of the predicates are available ahead of time. Furthermore, recent work using cheap approximations (i.e., proxies) do not support aggregation queries with predicates. To accelerate aggregation queries with expensive predicates, we develop and analyze a query processing algorithm that leverages proxies (ABAE). ABAE must account for the key challenge that it may sample records that do not satisfy the predicate. To address this challenge, we first use the proxy to group records into strata so that records satisfying the predicate are ideally grouped into few strata. Given these strata, ABAE uses pilot sampling and plugin estimates to sample according to the optimal allocation. We show that ABAE converges at an optimal rate in a novel analysis of stratified sampling with draws that may not satisfy the predicate. We further show that ABAE outperforms on baselines on six real-world datasets, reducing labeling costs by up to 2.3X.


2021 ◽  
Vol 46 (1) ◽  
pp. 1-46
Author(s):  
Jonas Traub ◽  
Philipp Marian Grulich ◽  
Alejandro Rodríguez Cuéllar ◽  
Sebastian Breß ◽  
Asterios Katsifodimos ◽  
...  

Window aggregation is a core operation in data stream processing. Existing aggregation techniques focus on reducing latency, eliminating redundant computations, or minimizing memory usage. However, each technique operates under different assumptions with respect to workload characteristics, such as properties of aggregation functions (e.g., invertible, associative), window types (e.g., sliding, sessions), windowing measures (e.g., time- or count-based), and stream (dis)order. In this article, we present Scotty , an efficient and general open-source operator for sliding-window aggregation in stream processing systems, such as Apache Flink, Apache Beam, Apache Samza, Apache Kafka, Apache Spark, and Apache Storm. One can easily extend Scotty with user-defined aggregation functions and window types. Scotty implements the concept of general stream slicing and derives workload characteristics from aggregation queries to improve performance without sacrificing its general applicability. We provide an in-depth view on the algorithms of the general stream slicing approach. Our experiments show that Scotty outperforms alternative solutions.


2021 ◽  
Vol 14 (6) ◽  
pp. 1033-1039
Author(s):  
James Thorne ◽  
Majid Yazdani ◽  
Marzieh Saeidi ◽  
Fabrizio Silvestri ◽  
Sebastian Riedel ◽  
...  

In recent years, neural networks have shown impressive performance gains on long-standing AI problems, such as answering queries from text and machine translation. These advances raise the question of whether neural nets can be used at the core of query processing to derive answers from facts, even when the facts are expressed in natural language. If so, it is conceivable that we could relax the fundamental assumption of database management, namely, that our data is represented as fields of a pre-defined schema. Furthermore, such technology would enable combining information from text, images, and structured data seamlessly. This paper introduces neural databases , a class of systems that use NLP transformers as localized answer derivation engines. We ground the vision in NeuralDB, a system for querying facts represented as short natural language sentences. We demonstrate that recent natural language processing models, specifically transformers, can answer select-project-join queries if they are given a set of relevant facts. However, they cannot scale to non-trivial databases nor answer set-based and aggregation queries. Based on these insights, we identify specific research challenges that are needed to build neural databases. Some of the challenges require drawing upon the rich literature in data management, and others pose new research opportunities to the NLP community. Finally, we show that with preliminary solutions, NeuralDB can already answer queries over thousands of sentences with very high accuracy.


Author(s):  
Tsz Nam Chan ◽  
Leong Hou U ◽  
Reynold Cheng ◽  
Man Lung Yiu ◽  
Shivansh Mittal

Author(s):  
Peeyush Gupta ◽  
Yin Li ◽  
Sharad Mehrotra ◽  
Nisha Panwar ◽  
Shantanu Sharma ◽  
...  

2019 ◽  
Vol 23 (1) ◽  
pp. 337-359
Author(s):  
Zheng Huo ◽  
Kerry Taylor ◽  
Xiuzhen Zhang ◽  
Suzhen Wang ◽  
Chaoyi Pang
Keyword(s):  

2019 ◽  
Vol 18 (1) ◽  
pp. 131-149
Author(s):  
Florian Johann Ganglberger ◽  
Joanna Kaczanowska ◽  
Wulf Haubensak ◽  
Katja Bühler

Sign in / Sign up

Export Citation Format

Share Document