Screening mammograms by community radiologists: variability in false-positive rates.

BACKGROUND Previous studies have shown that the agreement among radiologists interpreting a test set of mammograms is relatively low. However, data available from real-world settings are sparse. We studied mammographic examination interpretations by radiologists practicing in a community setting and evaluated whether the variability in false-positive rates could be explained by patient, radiologist, and/or testing characteristics. METHODS We used medical records on randomly selected women aged 40-69 years who had had at least one screening mammographic examination in a community setting between January 1, 1985, and June 30, 1993. Twenty-four radiologists interpreted 8734 screening mammograms from 2169 women. Hierarchical logistic regression models were used to examine the impact of patient, radiologist, and testing characteristics. All statistical tests were two-sided. RESULTS Radiologists varied widely in mammographic examination interpretations, with a mass noted in 0%-7.9%, calcification in 0%-21.3%, and fibrocystic changes in 1.6%-27.8% of mammograms read. False-positive rates ranged from 2.6% to 15.9%. Younger and more recently trained radiologists had higher false-positive rates. Adjustment for patient, radiologist, and testing characteristics narrowed the range of false-positive rates to 3.5%-7.9%. If a woman went to two randomly selected radiologists, her odds, after adjustment, of having a false-positive reading would be 1.5 times greater for the radiologist at higher risk of a false-positive reading, compared with the radiologist at lowest risk (95% highest posterior density interval [similar to a confidence interval] = 1.17 to 2.08). CONCLUSION Community radiologists varied widely in their false-positive rates in screening mammograms; this variability range was reduced by half, but not eliminated, after statistical adjustment for patient, radiologist, and testing characteristics. These characteristics need to be considered when evaluating false-positive rates in community mammographic examination screening.

[1]  N Segnan,et al.  Inter-observer and intra-observer variability of mammogram interpretation: a field study. , 1992, European journal of cancer.

[2]  B. Monsees,et al.  The Mammography Quality Standards Act. An overview of the regulations and guidance. , 2000, Radiologic clinics of North America.

[3]  K. Kerlikowske,et al.  Likelihood ratios for modern screening mammography. Risk of breast cancer based on age and mammographic interpretation. , 1996, JAMA.

[4]  S. Taplin,et al.  Effect of estrogen replacement therapy on the specificity and sensitivity of screening mammography. , 1996, Journal of the National Cancer Institute.

[5]  L. Liberman,et al.  Breast imaging reporting and data system (BI-RADS). , 2002, Radiologic clinics of North America.

[6]  K. Larsen,et al.  Interpreting Parameters in the Logistic Regression Model with Random Effects , 2000, Biometrics.

[7]  K. Kerlikowske,et al.  Variability and accuracy in mammographic interpretation using the American College of Radiology Breast Imaging Reporting and Data System. , 1998, Journal of the National Cancer Institute.

[8]  J. Elmore,et al.  Variability in radiologists' interpretations of mammograms. , 1994, The New England journal of medicine.

[9]  E A Sickles,et al.  Quality assurance. How to audit your own mammography practice. , 1992, Radiologic clinics of North America.

[10]  C L Robertson,et al.  A private breast imaging practice: medical audit of 25,788 screening and 1,077 diagnostic examinations. , 1993, Radiology.

[11]  C J Baines,et al.  The role of the reference radiologist. Estimates of inter-observer agreement and potential delay in cancer detection in the national breast screening study. , 1990, Investigative radiology.

[12]  S J Schnitt,et al.  Interobserver Reproducibility in the Diagnosis of Ductal Proliferative Breast Lesions Using Standardized Criteria , 1992, The American journal of surgical pathology.

[13]  C. Rutter,et al.  Assessing mammographers' accuracy. A comparison of clinical and test performance. , 2000, Journal of clinical epidemiology.

[14]  S L Zeger,et al.  Multivariate Continuation Ratio Models: Connections and Caveats , 2000, Biometrics.

[15]  Craig A. Beam,et al.  Variability in the interpretation of screening mammograms by US radiologists. Findings from a national sample. , 1996, Archives of internal medicine.

[16]  T. Eberlein,et al.  Biopsy of occult breast lesions. Analysis of 1261 abnormalities. , 1990, JAMA.

[17]  E A Sickles,et al.  Initial versus subsequent screening mammography: comparison of findings and their prognostic significance. , 1995, AJR. American journal of roentgenology.

[18]  G O Barnett,et al.  The application of computer-based medical-record systems in ambulatory practice. , 1984, The New England journal of medicine.

[19]  R. J. Brenner,et al.  Mammographic changes after excisional breast biopsy for benign disease. , 1996, AJR. American journal of roentgenology.

[20]  J. Elmore,et al.  A bibliography of publications on observer variability (final installment). , 1992, Journal of clinical epidemiology.

[21]  T. Eberlein,et al.  Biopsy of Occult Breast Lesions and Professional Liability-Reply , 1990 .

[22]  G.O. Barnett,et al.  COSTAR—A computer-based medical information system for ambulatory care , 1979, Proceedings of the IEEE.

[23]  Michael P. Jones Indicator and stratification methods for missing explanatory variables in multiple linear regression , 1996 .

[24]  R. Bird,et al.  Low-cost screening mammography: report on finances and review of 21,716 consecutive cases. , 1989, Radiology.

[25]  P. Rheinstein,et al.  The Mammography Quality Standards Act of 1992. , 1994, American family physician.

[26]  C. Beam,et al.  Variability in the interpretation of screening mammograms by US radiologists. Findings from a national sample. , 1996, Archives of internal medicine.

[27]  R. Smith,et al.  The mammography audit: a primer for the mammography quality standards act (MQSA). , 1995, AJR. American journal of roentgenology.

[28]  Jo M. Kendrick,et al.  Quality Determinants of Mammography, Clinical Practice Guideline , 1995 .

[29]  C J Baines,et al.  Menstrual cycle variation in mammographic breast density: so who cares? , 1998, Journal of the National Cancer Institute.

[30]  P. Vineis,et al.  Inter-Observer Variability in the Interpretation of Mammograms , 1988, Tumori.

[31]  P. Diggle Analysis of Longitudinal Data , 1995 .

[32]  E A Sickles,et al.  Screening mammography in community practice: positive predictive value of abnormal findings and yield of follow-up diagnostic procedures. , 1995, AJR. American journal of roentgenology.

[33]  A R Feinstein,et al.  A bibliography of publications on observer variability. , 1985, Journal of chronic diseases.

[34]  A. Gelfand,et al.  Predicting the cumulative risk of false-positive mammograms. , 2000, Journal of the National Cancer Institute.

[35]  J. Elmore,et al.  Ten-year risk of false positive screening mammograms and clinical breast examinations. , 1998, The New England journal of medicine.

[36]  W W Hauck,et al.  A consequence of omitted covariates when estimating odds ratios. , 1991, Journal of clinical epidemiology.

[37]  Roger W. Jelliffe,et al.  COSTAR-A Computer-Based Medical Information System for Ambulatory Care , 1979 .

[38]  L. Tabár,et al.  Significant reduction in advanced breast cancer. Results of the first seven years of mammography screening in Kopparberg, Sweden. , 1985, Diagnostic imaging in clinical medicine.

[39]  F. Houn,et al.  The Mammography Quality Standards Act of 1992. History and philosophy. , 1995, Radiologic clinics of North America.

[40]  W Vach,et al.  Biased estimation of the odds ratio in case-control studies due to the use of ad hoc methods of correcting for missing values for confounding variables. , 1991, American journal of epidemiology.

[41]  Marydale Debor,et al.  Quality determinants of mammography , 1993 .