The paper is concerned with the problem of automatic detection and correction of erroneous data into large datasets. The adopted process should therefore be computationally efficient. As usual, errors axe here defined by using a rule-based approach: all and only the data records respecting a set of rules axe declared correct. Erroneous records should afterwards be corrected, by using as much as possible the correct information contained in them. By encoding such problem into propositional logic, for each erroneous record we have a propositional logic formula, for which we want a model having particular properties. Correction problems can therefore be quickly solved by means of a customized SAT solver. Techniques for an efficient encoding of...
The amount of data being collected is growing exponentially, both in academics as well as in busines...
Much effort is spent everyday by programmers in trying to reduce long, failing execution traces to t...
The Fellegi-Holt method automatically "corrects" data that fail some predefined requiremen...
The paper is concerned with the problem of automatic detection and correction of inconsistent or out...
In a variety of relevant real world problems, tasks of "data mining" and "knowledge discovery" are r...
The paper is concerned with the problem of automatic detection and correction of errors into massive...
The paper is concerned with the problem of automatic detection and correction of errors into massive...
Encoding Information Reconstruction problems of different origins into Propositional Satisfiability ...
One of the main challenges that data cleaning systems face is to automatically identify and repair d...
Data collected by statistical offices generally contain errors, which have to be corrected before re...
The aim of our work is to investigate how a relatively small set of clauses can be transformed into...
We study the problem of introducing errors into clean databases for the purpose of benchmarking data...
We study the problem of introducing errors into clean data-bases for the purpose of benchmarking dat...
Software applications have become an indispensable integral part of this world. In all areas of ever...
Error detection is key for data quality management. Leveraging domain knowledge in the form of user-...
The amount of data being collected is growing exponentially, both in academics as well as in busines...
Much effort is spent everyday by programmers in trying to reduce long, failing execution traces to t...
The Fellegi-Holt method automatically "corrects" data that fail some predefined requiremen...
The paper is concerned with the problem of automatic detection and correction of inconsistent or out...
In a variety of relevant real world problems, tasks of "data mining" and "knowledge discovery" are r...
The paper is concerned with the problem of automatic detection and correction of errors into massive...
The paper is concerned with the problem of automatic detection and correction of errors into massive...
Encoding Information Reconstruction problems of different origins into Propositional Satisfiability ...
One of the main challenges that data cleaning systems face is to automatically identify and repair d...
Data collected by statistical offices generally contain errors, which have to be corrected before re...
The aim of our work is to investigate how a relatively small set of clauses can be transformed into...
We study the problem of introducing errors into clean databases for the purpose of benchmarking data...
We study the problem of introducing errors into clean data-bases for the purpose of benchmarking dat...
Software applications have become an indispensable integral part of this world. In all areas of ever...
Error detection is key for data quality management. Leveraging domain knowledge in the form of user-...
The amount of data being collected is growing exponentially, both in academics as well as in busines...
Much effort is spent everyday by programmers in trying to reduce long, failing execution traces to t...
The Fellegi-Holt method automatically "corrects" data that fail some predefined requiremen...