Usability and hardcopy manuals: evaluating research designs and methods

For the past decade, testing the usability of print software manuals has become a mature area of study, characterized by a wide range of qualitative and quantitative methods. Some of the most common methods include field observations, surveys, interviews, protocol analyses, focus groups, iterative testing, and quasi-experimental lab simulations [1]. Such diverse approaches to usability testing offer an opportunity for complementary inquiries and analyses. For example, findings from focus groups can provide key questions for experimental researchers to pursue in greater depth and with greater possibility for generalizability. Essentially, this complementary approach envisions an interaction between the academy, with its propensity toward pure, experimental research, and industry, with its more applied approaches for alpha and beta testing. Patricia Wright, a specialist in usability studies, has long argued that integrating pure and applied research is the best means for expanding our knowledge about effective document design [2; 3]. Such integration reveals both the immediately applicable aspects of effective manuals and the more theoretical boundaries in textual features that make a difference for general types of tasks, readers, and contexts of use. In order to realize the potential of conducting a conversation between pure and applied research, documentation researchers and practitioners must clearly understand the limitations that exist in the conclusions that investigators derive from specific methods of inquiry. In this article, I look solely at experimental usability tests that rely on quantitative methods of analysis. I analyze the ways in which the research designs and questions of the past ten years of experimental studies affect the strength of cumulative conclusions and the confidence we can have in those conclusions. My purpose is not to give preference to experimental research as the most important approach to usability testing. Far from it. Rather my critical review has two purposes: (1) to facilitate the dialogue between academic and industrial researchers by identifying the limits of current experimental findings; and (2) to propose research agendas and designs for future experimental usability tests that can strengthen the conclusions that such researchers offer for practical consideration. My evaluation of ten years of experimental usability studies shows that many of the conclusions of these studies are not strong enough to serve as valid, generalizable, and replicable foundations for subsequent research, be it pure or applied. These conclusions can be strengthened by designing studies that pay more attention to the sequencing and integration of related investigations and that institute better controls for sample selection, size, and composition. This article discusses my overall findings, the details of which I will develop more fully in my presentation.

[1]  John M. Carroll,et al.  The Minimal Manual , 1987, SGCH.

[2]  Mary Beth Rosson,et al.  On Comprehending a Computer Manual: Analysis of Variables Affecting Performance , 1987, Int. J. Man Mach. Stud..

[3]  John M. Carroll,et al.  Training wheels in a user interface , 1984, CACM.

[4]  Davida Charney,et al.  Designing Interactive Tutorials for Computer Users , 1987 .

[5]  Gregory S. Krohn Flowcharts Used for Procedural Instructions , 1983 .

[6]  Richard E. Mayer,et al.  Instructional Manipulation of Users' Mental Models for Electronic Calculators , 1984, Int. J. Man Mach. Stud..

[7]  David E. Kieras,et al.  The Acquisition of Procedures from Text: A Production-System Analysis of Transfer of Training. Technical Report No. 16. , 1985 .

[8]  John M. Carroll,et al.  Exploring Exploring a Word Processor , 1985, Hum. Comput. Interact..

[9]  Patricia Wright,et al.  Feeding the information eaters: Suggestions for integrating pure and applied research on language comprehension , 1978 .

[10]  L M Reder,et al.  The role of elaborations in learning a skill from an instructional text , 1985, Memory & cognition.

[11]  John M. Carroll,et al.  Blocking Learner Error States in a Training-Wheels System , 1984 .

[12]  D. A. Schell,et al.  Testing online and print user documentation , 1986, IEEE Transactions on Professional Communication.

[13]  Kirk H. Smith,et al.  Training by Exploration: Facilitating the Transfer of Procedural Knowledge Through Analogical Reasoning , 1986, Int. J. Man Mach. Stud..

[14]  R. Mayer,et al.  Difference recall protocols for technical texts due to advance organizers. , 1980 .

[15]  R. Mayer,et al.  Difference recall protocols for technical texts due to advance organizers. , 1980 .

[16]  Thomas M. Duffy,et al.  Document Design for Technical Job Tasks: An Evaluation , 1983 .

[17]  P. Dixon The processing of organizational and component step information in written directions , 1987 .

[18]  Alan C. Schultz,et al.  Multilevel Structured Documentation , 1989 .

[19]  Cecilia Katzeff Dealing with a Database Query Language in a New Situation , 1986, Int. J. Man Mach. Stud..

[20]  E. K. Odescalchi Documentation is the key to user success , 1986, IEEE Transactions on Professional Communication.

[21]  P. Dixon Plans and Written Directions for Complex Tasks. , 1982 .

[22]  Jaap Vermeulen Effects of Functionally or Topographically Presented Process Schemes on Operator Performance , 1987 .

[23]  Thomas P. Moran,et al.  Mental models and problem solving in using a calculator , 1983, CHI '83.

[24]  Davida Charney,et al.  Initial Skill Learning: An Analysis of How Elaborations Facilitate the Three Components. , 1986 .

[25]  Richard J. Light,et al.  Six evaluation issues that synthesis can resolve better than single studies , 1984 .