Publication
Title
Improving data quality : consistency and accuracy
Author
Abstract
Two central criteria for data quality are consistency and accuracy. Inconsistencies and errors in a database often emerge as violations of integrity constraints. Given a dirty database D, one needs automated methods to make it consistent, i.e., find a repair D0 that satisfies the constraints and minimally differs from D. Equally important is to ensure that the automatically-generated repair D0 is accurate, or makes sense, i.e., D0 differs from the correct data within a predefined bound. This paper studies effective methods for improving both data consistency and accuracy. We employ a class of conditional functional dependencies (CFDs) proposed in [6] to specify the consistency of the data, which are able to capture inconsistencies and errors beyond what their traditional counterparts can catch. To improve the consistency of the data, we propose two algorithms: one for automatically computing a repair D0 that satisfies a given set of CFDs, and the other for incrementally finding a repair in response to updates to a clean database. We show that both problems are intractable. Although our algorithms are necessarily heuristic, we experimentally verify that the methods are effective and efficient. Moreover, we develop a statistical method that guarantees that the repairs found by the algorithms are accurate above a predefined rate without incurring excessive user interaction.
Language
English
Source (book)
Proceedings of the 33rd International Conference on Very Large Data Bases (VLDB 2007), University of Vienna, Austria, September 23-27, 2007 / Koch, Christoph [edit.]; et al.
Publication
S.l. : ACM , 2007
ISBN
978-1-59593-649-3
Volume/pages
p. 315-326
UAntwerpen
Faculty/Department
Publication type
Subject
External links
Record
Identifier
Creation 14.01.2014
Last edited 22.08.2023
To cite this reference