Running Head: Analyzing A Novel Expertise Analyzing a Novel Expertise: An Unmarked Road Wayne D. Gray George Mason University & Susan S. Kirschenbaum Naval Undersea Warfare Center Division Newport Gray, W. D., & Kirschenbaum, S. S. (in press). Analyzing a novel expertise: An unmarked road. In J. M. C. Schraagen, S. F. Chipman, & V. L. Shalin (Eds.), Cognitive task analysis . Mahwah, NJ: Erlbaum. Please send all correspondence to: Wayne D. Gray George Mason University MSN 3f5 Fairfax, VA 22030 1-703-993-1357 gray@gmu.edu Gray & Kirschenbaum Analyzing a Novel Expertise page 2 Analyzing a Novel Expertise: An Unmarked Road There are many varieties of task analysis-each with its advantages and disadvantages, each with its adherents and detractors (e.g. see the recent collections published by Annett & Stanton, 1998; Kirwan & Ainsworth, 1992). Most published descriptions focus on how to apply the technique or why it is a good technique to apply. Few accounts written by advocates of a technique are specifically directed at problems and pitfalls in applying the technique. This account is different. Although we are unabashedly enthusiastic advocates of the theory-driven combination of task analysis and protocol analysis that we employ, we hope that by identifying problems and obstacles that we encountered that more people will be better prepared and, therefore, more successful at applying these techniques. Beware—knowing that the road is narrow, winding, and unmarked does not make the trip easy. It might, however, discourage someone from setting out in the family sedan. For those who are better equipped, knowledge of the hazards ahead may help them avoid blindly plunging forward into a known problem. It is in this spirit that we write this chapter. The following section provides a brief overview of the techniques we employ. It then introduces the known obstacles to these techniques. The main part of the chapter discusses these obstacles in the context of a specific project—Project Nemo. Theory-Driven Task Analysis and Protocol Analysis Theory-driven task analysis decomposes the procedural and declarative knowledge required to perform a task into components supported by the theory. With some additional work on the part of the analyst, the control structure provided by the theory can use the elements of the analysis Gray & Kirschenbaum Analyzing a Novel Expertise page 3 to form a model of how a user performs the task. Theories with weak or rigid control structures, such as keystroke-level GOMS or CPM-GOMS (for an overview, see John & Kieras, 1996a; John & Kieras, 1996b), may produce models that are only capable of performing the exact task that was analyzed. Theories with more powerful control structures, such as NGOMSL, ACT-R, Soar, or EPIC (see Gray, Young, & Kirschenbaum, 1997b), may respond adaptively to perform variations of the analyzed task. Cognitive theories provide constraints to the final form of the analysis—that is, for how the components must fit together. However, the components per se vary widely, and the analysis of expertise into such components is an underconstrained problem. For example, the expertise exhibited by a chess master in plotting his next move is different from that shown by a medical expert diagnosing a rare disease (VanLehn, 1989). Once the components (i.e., the knowledge structures and strategies) of expertise have been delineated, they can be cast into the mechanisms of a cognitive theory. However, existing cognitive theory provides few a priori constraints for deriving the components of a hitherto unstudied expertise. Given a rare form of expertise, one that has been subject to few published reports (Kirschenbaum, 1990; 1992; 1994), how does the analyst proceed? The method adopted here is a form of bootstrapping. As shown in Figure 1, an initial task analysis1 guides a protocol analysis. The task analysis is revised and used as the basis of the next round of protocol analysis. When the analyst deems that the results of the analysis are as good as the existing data permit, the analyst moves to the next phase of the effort. Gray & Kirschenbaum Analyzing a Novel Expertise page 4 Initial Task Analysis Protocol Analysis Results of Analysis As good as the data permit? No Revise Task Analysis Yes Next Phase Figure 1: Bootstrapping The story told in this chapter is the story of an iterative loop around the stages shown in Figure 1. The project has now moved to the next phase. The story of the next phase is ongoing; its outlines are given in Ehret, Kirschenbaum, and Gray (1998) and Ehret, Gray, and Kirschenbaum (1999). Dead Ends and Wrong Turns Studying an expertise that has not been extensively analyzed is like traveling on an unmarked Gray & Kirschenbaum Analyzing a Novel Expertise page 5 road – one with many intersections and forks. What aspects of the expertise are important for the goals of the analysis? Can knowledge of the components of other expertise guide and inform the current analysis or does this knowledge serve to lead us astray? If the expertise is a dynamic expertise – problem solving that takes place over a period of time and that is responsive to events in the environment – what is the best way to capture key aspects of the expertise without distorting it due to artifacts of data collection? These problems are common in analyzing expertise. Indeed, others have warned that an almost inevitable danger of doing the first, deep-level cognitive analysis of a hitherto unstudied expertise is that “the final interpretation of the data and its matching against the theory [i.e., task analysis] may appear to reflect mainly hindsight – pulling out of the data the hypotheses that have been implicitly hidden in it by the data interpretation process” (Ericsson & Simon, 1993, p. 286). Although we have read (and heeded) the warnings, what we have not read is a detailed discussion by analysts of their encounters with problems that threaten the validity of their conclusions. This chapter fills that void. In it, we provide an autobiographic description of the problems encountered in our analyses for Project Nemo. We feel no shame at admitting to having problems. Indeed, when the road is rocky and unmarked, problems must be expected. The shame lies not in having problems, but in not recognizing problems. The shame falls to the analyst who mistakes a dead end for the end of the trip or a wrong turn for the right path. The Task and Our Goals in Analyzing It Different cognitive task analyses may have different goals. The goal of Project Nemo was to analyze the knowledge and cognitive processes used by submarine approach officers (AOs) as Gray & Kirschenbaum Analyzing a Novel Expertise page 6 they tried to localize an enemy submarine hiding in deep waters. The project is a collaboration between a navy and a university researcher. An important role played by the navy researcher was to feed the results of the project, as they came in, to those parts of the submarine research community that could make the most use of them. From the beginning of the project, the most interested parties have been the designers of the command workstation for the next-generation submarine. Knowing that the results of the analysis would be used for interface design provided an important constraint on the knowledge and cognitive processes that Project Nemo analyzed. The AO possesses specialized knowledge that is acquired over a 20-year period. Rather than focus on the acquisition, depth, and breadth of this knowledge, we focused on how it is used as the AO makes progress in his goal of localizing the enemy submarine. Our effort focused on the knowledge, structures, and strategies unique to the dynamic problem-solving process of localizing an enemy submarine hiding in deep water.2 Any given task takes place in the context of the artifacts and organization used to perform the task. The goal of many task analyses, including most GOMS analyses, is to analyze task interactions at this activity level. Such analyses can either assess the problems with the current way of performing the task or provide specifications for a new system. The goal of the current analysis is different. We are neither involved in critiquing the current system nor directly involved in designing the new system. Rather, our intent is to provide the designers of the new system with a detailed description of the information processing performed by the AO as he Gray & Kirschenbaum Analyzing a Novel Expertise page 7 localizes an enemy submarine. Hence, in our analysis, we seek to abstract from the activity level to the functional level (Gray, John, & Atwood, 1993, pp. 244-257). Issues We had two sets of problems in understanding AO expertise. The first set included issues in understanding the control structure of the cognitive processes used by the AO to perform the localizing task. The second set entailed understanding the nature and limits to the data we had collected. The Control Structure of AO Expertise A common metaphor is to conceptualize problem solving as search in a problem space (Newell & Simon, 1972). Tasks involving expertise are often thought of as being both wide and deep. At each step in the problem space, there are many alternative next steps (width). Solving a problem involves solving many subproblems, and each subproblem can be decomposed into another subproblem that needs to be solved (depth). In contrast to expertise, the accepted wisdom is that, for everyday tasks, the search space is limited. The problem spaces for everyday tasks are either shallow and wide (like choosing a flavor from the menu in an ice cream store) or narrow and deep (like following a recipe from a cookbook) (Norman, 1989). Clearly, localizing targets hiding in deep water is not an everyday activity. Therefore, we reasoned that rather than being shallow and wide or narrow and deep that the AOs’ expertise must be both wide and deep. This bias led us to make our first and most fundamental mistake regarding the control structure of the AOs’ cognitive processes for this task. Gray & Kirschenbaum Analyzing a Novel Expertise page 8 Matrix? Subgoals? One of the most basic control structures used in cognitive task analysis is hierarchical, or goalsubgoal, decomposition. From discussions with experts, however, it was clear that AOs kept track of many different pieces of several different tasks. Hence, a classic, hierarchical control structure did not seem accurate. Our alternative was a matrix goal structure. Although we abandoned the matrix notion before working out the details, its basic elements were as follows. Imagine a cube with AO goals along the x axis, information elements about the target, own ship, and ocean conditions along the y axis, and rules that capture the AOs’ procedural knowledge along the z axis. A given rule would yield information regarding one or more information element. In turn, a given information element would be applicable to one or more AO goals. In such a system, one could imagine that the rules that fired (i.e., the actions that the AOs take) would be those rules that yielded the greatest amount of new information for the greatest number of AO goals. Unfortunately, for our preconceptions, we found little in our data that would support a matrix organization. We gradually abandoned this idea and, for a while, ignored this issue to concentrate on the issue of how and why one rule rather than another is selected. Our explorations of these control structure issues led us, albeit a bit unwillingly, to the realization that most AO actions could be characterized as small steps in a shallow goal hierarchy. However, unlike the everyday task of choosing one flavor from a wide but shallow ice cream store menu, AOs make many successive choices. It is the nature of these successive choices that characterize the AOs’ procedural expertise. Gray & Kirschenbaum Analyzing a Novel Expertise page 9 Our current theory of how AOs solve the localizing problem can be summarized by the rather awkward phrase “schema-directed problem solving with shallow and adaptive subgoaling” (SDPSSAS). The schema is the task-relevant knowledge accumulated over 20 years of experience as a submariner (half of it at sea). It is a knowledge structure3 that contains both declarative and procedural knowledge. An implication of shallow subgoaling is that the knowledge available to AOs is so rich that the steps required to supplement this knowledge can be fairly shallow. The second implication is that the problem the AO is constantly solving is “what is the state of the world – NOW” (where NOW is somewhere on the order of 30 to 300 s). The AO is trying to find a quiet target hiding in a noisy environment while remaining covert and undetected himself. What we see him doing is taking short steps that either (a) assess the noise from the environment or signal from the target – NOW, or (b) attempt to reduce the noise or increase the signal from the target by maneuvering own ship. As shown in Figure 2, these short steps result in shallow subgoaling. When a subgoal pops, the schema is reassessed. The result of this reassessment directs the next step (i.e., selects the next subgoal). This step is accomplished, it returns information to the schema, the schema is reassessed, and so on The process of subgoaling is adaptive in two senses. First, the subgoal that is chosen next reflects the current reassessment of the schema. Second, this choice is sensitive to both the longterm importance of the subgoal as well as its recent history of success or failure. Regardless of a goal’s long-term importance, AOs will not continue to attempt a goal if successive tries fail. Instead, they will choose another goal and return to the more important goal later.4 Gray & Kirschenbaum Analyzing a Novel Expertise page 10 The dynamic aspect of the AO's task plays an important role in this view of schema-directed problem solving with shallow and adaptive subgoaling. First, the state of the AO's world is continually changing – both own ship and target are moving at a given depth, direction, and speed. For own ship, the value of these attributes can be changed, but neither own ship nor the target can be stopped. Consequently, time is an important part of the picture. Second, subgoals are not accomplished once and then discarded. In the AO's world, subgoals bring in certain types of information or accomplish certain changes to own ship. As the world changes, any given subgoal may be revisited (e.g., DET-BEARING in Figure 2). LOCATESUB (t) DETBEARING LOCATESUB (t+1) EVALSNR LOCATESUB (t+2) LOCATESUB (t+3) SETTRACKER LOCATESUB (t+4) DETBEARING LOCATESUB (t+n) TMA-SOL Time (t) Figure 2: Schema-Directed Problem-Solving with Shallow and Adaptive Subgoaling (SDPSSAS) Choosing What to Do Next: Goal-Driven? Event-Driven? When we put aside the issue of the larger organization of the AO's actions – subgoal or matrix (at this point, we had not hit on SDPSSAS) – we focused on the more local issue of how or why an individual rule was chosen. For problems such as arithmetic, physics problems, and Tower of Hanoi, the pacing of the solution is entirely under the control of the person solving the problem. The problem can be decomposed into a deep and increasingly wide series of goals and subgoals. Gray & Kirschenbaum Analyzing a Novel Expertise page 11 At any given place in this goal-subgoal hierarchy, the action chosen next completely depends on the plans and knowledge of the problem solver. Such tasks are said to be goal-driven. We recognized from the beginning that the AO’s problem is not simply goal-driven in that the problem state changes without the intervention of the person. Some part of what the AO does is clearly an event-driven process. The issue is not event-driven versus goal-driven, but what combination of the two controls the AO's problem-solving behavior and how this combination can be represented in a cognitively plausible manner. As we were struggling with the goal-driven versus event-driven interpretations, we were having problems supporting another early idea – localizing a target as a diagnostic process. The literature suggests that an important first stage in solving many tasks is diagnosing the problem. This stage ends when the correct schema is selected (VanLehn, 1989). We fully expected that the first stage of localizing the target would entail diagnosing the situation to determine the correct target-finding schema. Indeed, for about the first 18 months, our encodings had a goal called IDENTIFY-POSSIBLE-INITIAL-SCHEMA. Believing that the “absence of evidence is not evidence of absence,” we struggled mightily to find support for a diagnostic process. However, we could not find evidence in our protocols or elsewhere (e.g., from any of the other AOs and experts whom we consulted) that we could interpret as evidence that schema selection required extended deliberation. Our tenacity in clinging to the notion of schema selection is, in part, attributable to evidence that suggested that AOs used different schemas when the target was a submarine than when the target was a merchant. Unfortunately, this evidence turned out to be bogus. Gray & Kirschenbaum Analyzing a Novel Expertise page 12 In brief, our analyses indicate that when LOCATE-MERCHANT was the goal, fewer subgoals were pushed and popped than when LOCATE-SUBMARINE was the goal (see the left side of Figure 3). More important, most of the subgoals for LOCATE-MERCHANT were similar to, but qualitatively different from, those for LOCATE-SUBMARINE. As discussed later, what turned out to be bogus was the apparent qualitative difference between similar subgoals for different goals. Once this spurious qualitative difference was eliminated, we concluded that the same LOCALIZE-THE-TARGET schema guided the processing of localizing either a hostile submarine or a friendly merchant. This conclusion converged with our failure to find any evidence of schema selection. We currently believe that schema selection is not a problem. Whether the goal is LOCATE-SUBMARINE or LOCATE-MERCHANT, our expert AOs have one schema, LOCALIZE-THE-TARGET, that is automatically chosen. Schema-directed Problem Solving Conceptualizing the AOs’ task as schema instantiation, not selection, was the key to a parsimonious account of the general structure of the AOs’ cognitive processes. Localizing the target was not driven by goals or events, but was directed by a schema. The data gathered were used to instantiate attributes of this schema. On each cycle of problem solving, the schema was reevaluated, the currently most critical attribute-value pair was identified, and instantiating this pair became the current subgoal. New events were incorporated into the schema and affect problem solving only to the extent that they affect the identification of the currently most critical attribute-value pair. Gray & Kirschenbaum Analyzing a Novel Expertise page 13 subgoal per goal operators per subgoal 8 number 6 4 2 0 LOC-MERC LOC-SUB LOC-MERC LOC-SUB Figure 3: Mean number of subgoals per goal (left two) and mean number of operators per subgoal (right two) for LOCALIZE-MERC and LOCALIZE-SUB. Also shown are the 95% confidence intervals for the standard error of the mean. This view of schema instantiation as schema-directed problem solving fits in nicely with our emerging awareness of shallow subgoaling. Each time the AO returned to the top-level goal (LOCATE-SUBMARINE in Figure 2), the schema was reevaluated. The subgoal chosen was one that would return information regarding one attribute-value pair. Hence, localizing a target is a wide and shallow task. The width is represented by a well-learned schema. The shallowness is represented by shallow subgoaling. Localizing the target involves dozens of iterations of schema evaluation and shallow subgoaling. These iterations continue until the AO has confidence that the schema as currently instantiated provides accurate information regarding the target’s location. Gray & Kirschenbaum Analyzing a Novel Expertise page 14 Data Analysis: Levels of Analysis and Limits to the Data Changes in our taxonomy of goals, subgoals, and operators were driven by three sources: our changing understanding of the general structure of AO cognition (discussed earlier), our failure to understand the limits to our data (discussed next), and our tendency to embed specifics of interacting with the simulation in our analysis of the task – that is, a failure to distinguish between task and artifact (discussed later). Distinctions Not Supported Our final encodings resulted in nine categories of operators (Gray, Kirschenbaum, & Ehret, 1997a; Kirschenbaum, Gray, & Ehret, 1997). Each of these categories can be considered as representing many subcategories. For example, at one time, we tried to distinguish subcategories of things that could be queried and sources from which information could be received. An AO could query a display, the own ship operator5, his own memory, an instrument reading, and so on. Likewise, information could be received from long-term memory, from short-term memory, from reading a table, from viewing a graphic, from inferring a relationship among other information, from the own ship operator, and so on. At one time or another, each of these subcategories was considered – if only briefly – as a candidate for inclusion in the analysis. In deriving our encoding categories, we focused on two transcripts, one from each of two AOs. New encoding categories could be proposed by any of our three encoders. Typically, the encoder would present a rationale for why the subcategory was needed, along with a particular instance from the transcript that the encoder felt represented that subcategory. If the other two encoders were convinced by the instance, much discussion would be spent defining criteria by which other instances could be identified. Two or three of the encoders would together go through one of the Gray & Kirschenbaum Analyzing a Novel Expertise page 15 two sample transcripts. New instances of the encoding category would be identified and debated until consensus was achieved. After this, two or three of the encoders would independently go through the other sample transcript in an attempt to identify instances of the new category. More often than not, each independent encoding would yield a handful of instances (out of about 300 encoded operators) that could be considered members of the new category. However, there would be little or no agreement about which encodings were exemplars of the new category and which were not. At this point, the candidate would be abandoned. We considered it a plausible distinction, but a distinction for which the data were at too coarse of a level to support. Problem Solving the Tool as Opposed to Problem Solving the Task We collected data from AOs who localized targets presented on a dynamic simulation of the ocean environment – CSEAL (for more information, see Gray et al., 1997a; or Kirschenbaum et al., 1997). However, our task was not to describe how AOs’ used the simulation (the artifact or tool) to localize the target (the task), but to separate the specifics of the simulation from the more general aspects of their problem-solving process (i.e., a functional level of analysis). Time and again, we found ourselves encoding specifics of using the simulator to localize targets rather than simply localizing targets. We attempted to encode every segment of the transcript, but realized early on that a significant number of segments simply had nothing to do with the problem. The fact that many of the segments in a verbal protocol have nothing to do with the problem being solved is not news and, indeed, we expected this. These NAs (not applicable) were easy to identify because they represented comments about the desirability of eating lunch soon, the drive to the base, the Gray & Kirschenbaum Analyzing a Novel Expertise page 16 weather, the room in which the simulation was run, and so on. Of the 2,882 segments encoded, 946 were in this NA category. More difficult to distinguish were the segments that did not refer to solving the problem but to some aspect of the simulation. An analogy to this is talking to a co-author about how to format a table in Microsoft Word™ as opposed to discussing the data that would go into the table. A simple example is represented by the following query from the AO to the own ship operator asking about the units (relative or true) in which a particular display represented the data. Protocol: “and so this [display] is relative or true? (AO points to TB16 block)?” At some point, we realized that what we were seeing was a ubiquitous phenomenon that was not limited to our paradigm, but that represented a type of usability issue. We saw this as problem solving in the tool space, as opposed to the task space, and proposed the tool:task ratio as a usability metric (Kirschenbaum, Gray, Ehret & Miller, 1996). Following this insight, our transcripts were recoded with a new operator, instrumentation. As per the prior example, most of the instrumentation operators were found in groups of one to three amid a series of task operators. These small groups of instrumentation operators are essentially asides that are embedded among operators concerned with a particular task goal. Occasionally, both the AO and the own ship operator abandoned the task of localizing the target and became engaged in an episode of collaborative problem solving – attempting to figure out how to get the simulation to take a particular input or to display a particular type of data. Such episodes were recoded as a new goal – that of supervising the use of the tool (supervisory). Gray & Kirschenbaum Analyzing a Novel Expertise page 17 Over our entire set of encoded data, 421 goals and 2,882 operators were required to encode the nine scenarios from our six AOs (for more details, see Gray, Kirschenbaum, et al., 1997 or Kirschenbaum et al., 1997). After removing all supervisory goals, their operators, all instrumentation operators, and all NA operators, we were left with 397 goals and 1,269 operators. We refer to this remainder as our clean set. The clean set of encodings was used in all subsequent analyses. The significance of this reduction cannot be overstated; over half of our encoded utterances had nothing to do with localizing per se. Having these in the analyses confounded our efforts to make sense of the data. Once these were removed, regularities that had been obscured became apparent. For example, the scenarios we studied involved two targets: a hostile submarine and a friendly merchant. Given the shallow subgoaling we were beginning to believe in, it made sense to us that there would be more subgoals involved in localizing the quiet submarine than in localizing the noisy merchant. Indeed, this is what our encodings suggested (see the left two data points in Figure 3). However, the same subgoal seemed to involve many more steps if its supergoal was LOCATE-SUBMARINE rather than LOCATE-MERCHANT. More steps were needed for a subgoal such as determine the signal-to-noise ratio (determineSNR) when determine-SNR was a subgoal of LOCATE-SUBMARINE than when it was a subgoal of LOCATE-MERCHANT, and this did not seem unreasonable. Indeed, as discussed earlier, such differences supported our belief that different schemas were used for different targets. By inference, this finding supported the belief that schema selection was an important component of the AO's problem-solving process. Gray & Kirschenbaum Analyzing a Novel Expertise page 18 However, once the instrumentation operators and supervisory goals were removed, regularities appeared. As shown in the right two data points of Figure 3, the subgoals used in LOCATEMERCHANT required the same number of operators as the subgoal of LOCATESUBMARINE. (An extended discussion of this point is provided in both Gray et al., 1997a; and Kirschenbaum et al., 1997.) Completed versus Successful Goals Another part of our struggle to define the number of levels and depth of the goal stack resulted from an implicit assumption that a completed goal was synonymous with a successful goal. In our initial attempts to shoehorn reality, we viewed a completed goal as one that returned the information queried. For example, if the AO queried the bearing rate on a particular target, this goal would not be completed until the target’s bearing rate was determined. Attempting to trace the path from initial query to completion led us to postulate a tangled web of semi-infinite subgoaling. Stepping back and listening to the data led to a different conclusion. We discovered that, for the AO, knowing that, at a given point in time under current conditions, the bearing rate (or course, speed, etc.) cannot be determined is an important and complete piece of information. The insight that a goal can be considered completed without being considered successful supports the shallow subgoaling component of schema-directed problem-solving. The AO launches a continuing stream of short queries. Each query returns some information. In the early stages of localizing (immediately after the target has been detected), the information typically is something such as the data are too noisy to answer that question (the SNR is literally too low). This leads the AO to take actions to increase the SNR. Gray & Kirschenbaum Analyzing a Novel Expertise page 19 Summary of Issues Understanding the control structure of the AOs’ cognition, together with understanding the nature and limits of the data, were major obstacles in our attempts to do a cognitive analysis of the AOs’ task. Before beginning this project, we knew that understanding the control structure of cognition would be the key to the cognitive task analysis (this concern was reflected in the original proposal to ONR). At each iteration around the loop (see Figure 1), each component of our analysis seemed plausible. What troubled us were our efforts to fit the parts together into a coherent whole. Although during this period we were not building ACT-R models, we were constantly asking ourselves how the disparate parts could fit into an ACT-R model. It was the failure to answer this question positively that kept driving us around the loop and deeper into the data. At this point, we have exited the loop (Figure 1) and moved onto the next phase of the project. Although the current hypothesis – schema-directed problem solving with shallow and adaptive subgoaling – is coherent, we are collecting additional data in the hopes of capturing finer grained data on key aspects of AO problem solving in a dynamic environment. Conclusions The current chapter has concentrated on the difficulties of doing a deep-level cognitive task analysis of a novel expertise. The difficulties are all the more notable in that our team of researchers brought to the study considerable expertise in cognitive theory and in applying cognitive theory to real-world tasks. Prior to working on Project Nemo, the co-authors of this chapter conducted research on COBOL programmers, HAWK Air Defense maintenance workers, Gray & Kirschenbaum Analyzing a Novel Expertise page 20 small-unit tactical team training, phone company operators, as well as more traditional decisionmaking studies of submariners and school children. Although the road was difficult and unmarked, we have arrived at our destination. Our current characterization of the AOs’ expertise – schema-directed problem solving with shallow and adaptive subgoaling – is both simpler and more profound than what we had envisioned when we began our journey. As far as we can tell, this characterization is unlike any that appears in the literature on expert performance. As such, it is important that those who are designing the command workstation understand this characterization of the AOs’ expertise rather than designing an interface that will support the consideration of multiple hypotheses (as in medical diagnoses) or the in-depth exploration of several alternative courses of action (as in chess playing). Instead of telling stories about how difficult our trip was, we would rather give the reader a sure-fire guide to plotting a safe path to any destination, on any road, marked or unmarked. We do not know if such a guide can be written. However, we are sure that we cannot write one. Unfortunately, the truth remains that, whatever may be done differently, the task of understanding a hitherto unstudied expertise will never be quick or easy. The problems discussed in this chapter can be anticipated but not avoided. Notes Acknowledgment More than usual, we thank our agency sponsors and our scientific officer for understanding that if we knew what we were doing, it would not be called research. We believe that the emerging Gray & Kirschenbaum Analyzing a Novel Expertise page 21 results justify their long-term support of this effort. However, we also understand that there were times when they might have thought that a successful outcome was unlikely. We thank Brian D. Ehret who joined our project about two thirds of the way through the events recounted here. Brian was the third encoder on each of the final encodings of the transcripts. His diagnoses have guided the current phase of data collection. The work on this project at George Mason University was supported by a grant from the Office of Naval Research (#N00014-95-1-0175) to Wayne D. Gray. Susan S. Kirschenbaum’s work has been jointly sponsored by Office of Naval Research (ONR) (Program element 61153N) and by Naval Undersea Warfare Center's Independent Research Program as Project A10328. Authors present address Wayne D. Gray can be reached at Human Factors and Applied Cognitive Program, George Mason University, MSN 3f5, Fairfax, VA 22030, USA. E-Mail: gray@gmu.edu. Susan S. Kirschenbaum can be reached at Naval Undersea Warfare Center Division Newport, 1176 Howell St., Code 2211, Building 1171/1, Newport, RI 02841-1708, USA. Email: kirschenbaumss@csd.npt.nuwc.navy.mil. End Notes 1 The initial task analysis can be as informal as it has to be, but should be as formal as possible. For Project Nemo, the initial task analysis was based on the published literature (Kirschenbaum, 1990; 1992; 1994). 2 After a target is detected, it must be localized. Detection tells the AO that a target is out there. Localizing tells him where it is in terms of bearing from own ship, range, course, and Gray & Kirschenbaum Analyzing a Novel Expertise page 22 speed. Due, in part, to the physics of sound transmission underwater and the need to remain covert, localizing a target is a mathematically underconstrained problem. Passive sonar is the only tool available to the AO. From passive sonar, the AO can directly compute the bearing of the target. Computing the target’s range, course, and speed is a difficult process. 3 As we think of the schema in terms of ACT-R mechanisms, the schema would be a body of task-specific, declarative memory elements and productions. Any given declarative memory element is relatively small and limited. However, the set of task-relevant, declarative memory elements have high interitem association values (see Anderson & Lebiére, 1998). 4 As described by Lovett (1998), this adaptive subgoaling can be modeled in ACT-R 4.0 as the temporary depression and recovery of the expected value of a goal. 5 The simulation required extensive training to operate. Rather than teaching AOs this esoteric task, the simulation was run by an experimenter in the role of own ship operator. This arrangement mimicked procedures onboard submarines and was acceptable to all of the AOs. References Anderson, J. R., & Lebiére, C. (Eds.). (1998). Atomic components of thought. Hillsdale, NJ: Erlbaum. Annett, J., & Stanton, N. (1998). Introduction to this Special Issue on Task Analysis. Ergonomics, 41(11), 1529-1536. Ehret, B. D., Gray, W. D., & Kirschenbaum, S. S. (1999). Scaled worlds and computational cognitive models as tools to increase the usefulness of cognitive task analysis. manuscript submitted for publication. Gray & Kirschenbaum Analyzing a Novel Expertise page 23 Ehret, B. D., Kirschenbaum, S. S., & Gray, W. D. (1998). Contending with complexity: the development and use of scaled worlds as research tools, Human Factors and Ergonomics Society 42th Annual Meeting . Santa Monica, CA: Human Factors and Ergonomics Society. Ericsson, K. A., & Simon, H. A. (1993). Protocol analysis: Verbal reports as data. (Revised ed.). Cambridge, MA: The MIT Press. Gray, W. D., John, B. E., & Atwood, M. E. (1993). Project Ernestine: Validating a GOMS analysis for predicting and explaining real-world performance. Human-Computer Interaction, 8(3), 237-309. Gray, W. D., Kirschenbaum, S. S., & Ehret, B. D. (1997a). The précis of Project Nemo, phase 1: Subgoaling and subschemas for submariners, Nineteenth Annual Conference of the Cognitive Science Society (pp. 283-288). Hillsdale, NJ: Erlbaum. Gray, W. D., Young, R. M., & Kirschenbaum, S. S. (1997b). Introduction to this Special Issue on Cognitive Architectures and Human-Computer Interaction. Human-Computer Interaction, 12(4), 301-309. John, B. E., & Kieras, D. E. (1996a). The GOMS family of user interface analysis techniques: Comparison and contrast. ACM Transactions on Computer-Human Interaction, 3(4), 320351. John, B. E., & Kieras, D. E. (1996b). Using GOMS for user interface design and evaluation: Which technique? ACM Transactions on Computer-Human Interaction, 3(4), 287-319. Kirschenbaum, S. S. (1990). Command Decision Making: Lessons Learned (NUSC TM No. Gray & Kirschenbaum Analyzing a Novel Expertise page 24 902149): Naval Underwater Systems Center. Kirschenbaum, S. S. (1992). The effect of level of experience on information gathering strategies. Journal of Applied Psychology, 77, 343-352. Kirschenbaum, S. S. (1994). Command Decision-Making: A Simulation Study (Technical Report 10,350): Naval Undersea Warfare Center Division Newport. Kirschenbaum, S. S., Gray, W. D., & Ehret, B. D. (1997). Subgoaling and subschemas for submariners: Cognitive models of situation assessment (Technical Report 10,764-1). Newport, RI: NUWC-NPT. Kirwan, B., & Ainsworth, L. K. (Eds.). (1992). A guide to task analysis. Washington, DC: Taylor & Francis. Lovett, M. (1998). Choice. In J. R. Anderson & C. Lebiére (Eds.), Atomic components of thought (pp. 255-296). Hillsdale, NJ: Erlbaum. Newell, A., & Simon, H. A. (1972). Human problem solving. Englewood Cliffs, NJ: PrenticeHall, Inc. Norman, D. A. (1989). The design of everyday things. New York: DoubleDay. VanLehn, K. (1989). Problem solving and cognitive skill acquisition. In M. I. Posner (Ed.), Foundations of Cognitive Science (pp. 527-579). Cambridge, MA: The MIT Press.