scholarly journals An R Script for Assessment of Data Quality in the BioSense Locker Database

2016 ◽  
Vol 8 (1) ◽  
Author(s):  
Serena Rezny ◽  
Stacey Hoferka

Syndromic surveillance requires reliable, accurate, and complete healthcare encounter data. To address the need for quality assessment of ED data, we developed an R script to assess and produce reports on data quality in the BioSense locker database. The script examines identifying variables in the HL7 messages from the locker, aggregates messages into ED visits based on these identifiers, processes the aggregated data to calculate metadata for each visit, and computes various data quality metrics. Facility-level reports are written to HTML files, which can then be shared with hospitals and vendors to support ongoing data quality improvements.

2008 ◽  
pp. 3067-3084
Author(s):  
John Talburt ◽  
Richard Wang ◽  
Kimberly Hess ◽  
Emily Kuo

This chapter introduces abstract algebra as a means of understanding and creating data quality metrics for entity resolution, the process in which records determined to represent the same real-world entity are successively located and merged. Entity resolution is a particular form of data mining that is foundational to a number of applications in both industry and government. Examples include commercial customer recognition systems and information sharing on “persons of interest” across federal intelligence agencies. Despite the importance of these applications, most of the data quality literature focuses on measuring the intrinsic quality of individual records than the quality of record grouping or integration. In this chapter, the authors describe current research into the creation and validation of quality metrics for entity resolution, primarily in the context of customer recognition systems. The approach is based on an algebraic view of the system as creating a partition of a set of entity records based on the indicative information for the entities in question. In this view, the relative quality of entity identification between two systems can be measured in terms of the similarity between the partitions they produce. The authors discuss the difficulty of applying statistical cluster analysis to this problem when the datasets are large and propose an alternative index suitable for these situations. They also report some preliminary experimental results, and outlines areas and approaches to further research in this area.


2018 ◽  
Vol 10 (1) ◽  
pp. 1-26 ◽  
Author(s):  
Christian Bors ◽  
Theresia Gschwandtner ◽  
Simone Kriglstein ◽  
Silvia Miksch ◽  
Margit Pohl

2011 ◽  
Vol 11 (2) ◽  
pp. 1412-1419 ◽  
Author(s):  
Christopher R. Kinsinger ◽  
James Apffel ◽  
Mark Baker ◽  
Xiaopeng Bian ◽  
Christoph H. Borchers ◽  
...  

2017 ◽  
Vol 20 (2) ◽  
Author(s):  
Flavia Serra ◽  
Adriana Marotta

The fact that Data Quality (DQ) depends on the context, in which data are produced, stored and used, is widely recognized in the research community. Data Warehouse Systems (DWS), whose main goal is to give support to decision making based on data, have had a huge growth in the last years, in research and industry. DQ in this kind of systems becomes essential. This work presents a proposal for identifying DQ problems in the domain of DWS, considering the different contexts that exist in each system component. This proposal may act as a first conceptual framework that guides the DQ-responsible in the management of DQ in DWS. The main contributions of this work are a thorough literature review about how contexts are used for evaluating DQ in DWS, and a proposal for assessing DQ in DWS through context-based DQ metrics.


2011 ◽  
Vol 10 (12) ◽  
pp. O111.015446 ◽  
Author(s):  
Christopher R. Kinsinger ◽  
James Apffel ◽  
Mark Baker ◽  
Xiaopeng Bian ◽  
Christoph H. Borchers ◽  
...  

2018 ◽  
Vol 10 (1) ◽  
Author(s):  
Peter J Rock ◽  
Michael D Singleton

Objective: The aim of this project was to develop a nimble system to both monitor and report on the quality of Kentucky emergency department syndromic surveillance (SyS) data at system-wide and facility levels.Introduction: In 2016, the CDC funded 12 states, under the Enhanced State Opioid Overdose Surveillance (ESOOS) program, to utilize SyS to increase timeliness of state data on drug overdose events. In order to operationalize the objectives of the grant, there was a need to assess and monitor the quality of Kentucky’s SyS data, with limited resources. We leveraged the NSSP’s R Studio Server to automate quality assurance (QA) monitoring and reporting to meet these objectives.Methods: Using the R Server, we pulled data from the process messages table, aggregating messages to single patient encounters. In addition to compiling the code on a powerful remote server, the server can access the process table messages relatively quickly. We developed an R Markdown report to produce a report that includes a variety of system- and facility-level metrics that highlight key indicators of system performance and data flows. By using R, we were able to create an auto-generating QA report that runs weekly and e-mails for analyst review. Quality metrics included: % completeness of chief complaint and discharge diagnosis codes (overall and by facility)[Fig 1 & Fig 2]; visit trend by day of visit (with interactive spark lines)[Fig 2]; maximum date of message created, date message arrived at NSSP server, date of visit, and total messages[Fig 3]; message arrived trend (interactive sparklines)[Fig 3]; volume and type of error messages failing to process[Fig 4]; message volume by ADT type[Fig 5]; and volume of patient class by type by day[not shown]. Our SyS analyst reviews the report and delivers it to stakeholders with general comments about ongoing and newly emerging data quality concerns.Results: The report has proven to be beneficial in ongoing QA monitoring. The report is shared weekly with key stakeholders at the Kentucky Department for Public Health, Kentucky Health Information Exchange, NSSP, and regional ESSENCE users. Findings are reviewed at monthly SyS stakeholder meetings. The report has identified numerous errors, dead feeds, and other systems changes in near real-time; leading to corrective action and general data quality enhancement. Weekly monitoring of QA has improved data feed stability and communication of identified issue with key stakeholders.Conclusions: The R Studio Server provides a nimble platform to develop, refine, and automate a QA reporting system that can lead to improved SyS data quality. In Kentucky, in addition to improving overall data quality, these weekly reports and subsequent communication have help built relationships among key stakeholders and elevated the importance of syndromic surveillance data locally. Continual monitoring of data is critical to ensure quality and therefor the validity of the data.


2018 ◽  
Vol 9 (2) ◽  
pp. 1-32 ◽  
Author(s):  
Bernd Heinrich ◽  
Diana Hristova ◽  
Mathias Klier ◽  
Alexander Schiller ◽  
Michael Szubartowicz

Sign in / Sign up

Export Citation Format

Share Document