Like other data quality dimensions, the concept of accuracy is often adopted to characterise a particular data set. However, its common specification basically refers to statistical properties of estimators, which can hardly be proved by means of a single survey at hand. This ambiguity can be resolved by assigning ‘accuracy’ to survey processes that are known to affect these properties. In this contribution, we consider the sub-process of imputation as one important step in setting up a data set and argue that the so called ‘hit-rate’ criterion, that is intended to measure the accuracy of a data set by some distance function of ‘true’ but unobserved and imputed values, is neither required nor desirable. In contrast, the so-called ‘inference’ criterion allows for valid inferences based on a suitably completed data set under rather general conditions. The underlying theoretical concepts are illustrated by means of a simulation study. It is emphasised that the same principal arguments apply to other survey processes that introduce uncertainty into an edited data set.
[1]
Jeremy MG Taylor,et al.
Partially parametric techniques for multiple imputation
,
1996
.
[2]
D. Rubin,et al.
Statistical Analysis with Missing Data
,
1988
.
[3]
D. Horvitz,et al.
A Generalization of Sampling Without Replacement from a Finite Universe
,
1952
.
[4]
D. Holt,et al.
A Systematic Approach to Automatic Edit and Imputation
,
1976
.
[5]
Carl-Erik Särndal,et al.
Model Assisted Survey Sampling
,
1997
.
[6]
D. Rubin.
Multiple Imputation After 18+ Years
,
1996
.
[7]
T. De Waal.
A Fast and Simple Algorithm for Automatic Editing of Mixed Data
,
2003
.
[8]
Paul P. Biemer,et al.
Introduction to Survey Quality
,
2003
.
[9]
Roger A. Sugden,et al.
Multiple Imputation for Nonresponse in Surveys
,
1988
.
[10]
Gordon Brackstone,et al.
Managing data quality in a statistical agency
,
2003
.