Reliability of Health Information on the Internet: An Examination of Experts' Ratings

Background The use of medical experts in rating the content of health-related sites on the Internet has flourished in recent years. In this research, it has been common practice to use a single medical expert to rate the content of the Web sites. In many cases, the expert has rated the Internet health information as poor, and even potentially dangerous. However, one problem with this approach is that there is no guarantee that other medical experts will rate the sites in a similar manner. Objectives The aim was to assess the reliability of medical experts' judgments of threads in an Internet newsgroup related to a common disease. A secondary aim was to show the limitations of commonly-used statistics for measuring reliability (eg, kappa). Method The participants in this study were 5 medical doctors, who worked in a specialist unit dedicated to the treatment of the disease. They each rated the information contained in newsgroup threads using a 6-point scale designed by the experts themselves. Their ratings were analyzed for reliability using a number of statistics: Cohen's kappa, gamma, Kendall's W, and Cronbach's alpha. Results Reliability was absent for ratings of questions, and low for ratings of responses. The various measures of reliability used gave conflicting results. No measure produced high reliability. Conclusions The medical experts showed a low agreement when rating the postings from the newsgroup. Hence, it is important to test inter-rater reliability in research assessing the accuracy and quality of health-related information on the Internet. A discussion of the different measures of agreement that could be used reveals that the choice of statistic can be problematic. It is therefore important to consider the assumptions underlying a measure of reliability before using it. Often, more than one measure will be needed for "triangulation" purposes.

[1]  M Pallen Guide to the Internet: Introducing the Internet , 1995, BMJ.

[2]  J. R. Landis,et al.  The measurement of observer agreement for categorical data. , 1977, Biometrics.

[3]  P. Whatling Having non-medical readers of papers on internet will enhance peer review , 1999, BMJ.

[4]  P. Impicciatore,et al.  Reliability of health information for the public on the world wide web: systematic survey of advice on managing fever in children at home , 1997, BMJ.

[5]  H. Sandvik Health information and interaction on the internet: a survey of female urinary incontinence , 1999, BMJ.

[6]  G Eysenbach,et al.  Information in practice Towards quality management of medical information on the internet : evaluation , labelling , and filtering of information , 1998 .

[7]  D H Gustafson,et al.  Consumers and evaluation of interactive health communication applications. The Science Panel on Interactive Communication and Health. , 1999, American journal of preventive medicine.

[8]  P. Kimmel,et al.  Poison on line--acute renal failure caused by oil of wormwood purchased through the Internet. , 1997, The New England journal of medicine.

[9]  J. Carlin,et al.  Bias, prevalence and kappa. , 1993, Journal of clinical epidemiology.

[10]  John Uebersax,et al.  Statistical Modeling of Expert Ratings on Medical Treatment Appropriateness , 1993 .

[11]  Gunther Eysenbach,et al.  Rapid Responses , 2022 .

[12]  W. Willett,et al.  Misinterpretation and misuse of the kappa statistic. , 1987, American journal of epidemiology.