Datasets:

leeclark commited on
Commit
01b54df
·
1 Parent(s): 7df4199

test: uploading a file via git

Browse files
20-F-1436-0-141 ADDED
File without changes
20-F-1436-1-141 ADDED
@@ -0,0 +1,141 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+
3
+
4
+
5
+
6
+
7
+
8
+
9
+
10
+
11
+
12
+
13
+
14
+
15
+
16
+
17
+
18
+
19
+
20
+
21
+
22
+
23
+
24
+
25
+
26
+
27
+
28
+
29
+
30
+
31
+
32
+
33
+
34
+
35
+
36
+
37
+
38
+
39
+
40
+
41
+
42
+
43
+
44
+
45
+
46
+
47
+
48
+
49
+
50
+
51
+
52
+
53
+
54
+
55
+
56
+
57
+
58
+
59
+
60
+
61
+
62
+
63
+
64
+
65
+
66
+
67
+
68
+
69
+
70
+
71
+
72
+
73
+
74
+
75
+
76
+
77
+
78
+
79
+
80
+
81
+
82
+
83
+
84
+
85
+
86
+
87
+
88
+
89
+
90
+
91
+
92
+
93
+
94
+
95
+
96
+
97
+
98
+
99
+
100
+
101
+
102
+
103
+
104
+
105
+
106
+
107
+
108
+
109
+
110
+
111
+
112
+
113
+
114
+
115
+
116
+
117
+
118
+
119
+
120
+
121
+
122
+
123
+
124
+
125
+
126
+
127
+
128
+
129
+
130
+
131
+
132
+
133
+
134
+
135
+
136
+
137
+
138
+
139
+
140
+
141
+
36733527-0-22 ADDED
File without changes
36733527-0-23 ADDED
File without changes
36733527-1-22 ADDED
@@ -0,0 +1,1274 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 2Monterey, Ca.tact identification, intent, available responses andtheir consequences. For example, the close prox-
2
+ imity of U.S. Navy forces and potential adversary
3
+ forces makes interpreting the actions of an in-
4
+ bound aircraft who does not respond to radio
5
+ warnings much more difficult. Should the air-
6
+ craft's behavior be interpreted as an attack profile,
7
+ or does the pilot merely intend to harass, or does
8
+ the aircraft in question not carry the equipment
9
+ necessary to receive verbal warnings, leaving the
10
+ pilot unable to receive radio warnings directed
11
+ toward him and unaware of his precarious posi-
12
+ tion? In extreme cases there is no clear cut right or
13
+ wrong answer about a decision. Rapidly unfolding
14
+ events result in severe time pressure and severe
15
+ (often catastro-phic) consequences for errors.
16
+ While current real-time battle management sys-
17
+ tems are well-suited to the demands of all-out
18
+ conflicts, they may not be optimized for littoral
19
+ situations where human intervention in decision-
20
+ making is even more important (Office of Naval
21
+ Technology [ONT], 1992). (Since 70 percent of
22
+ the world's population lives within 200 miles of
23
+ the sea, most future contingencies are likely to
24
+ involve littoral warfare (Mundy, 1994).)Two unfortunate and highly publicized eventsfocused attention on the difficult types of deci-
25
+ sions confronting naval commanders and provided
26
+ the impetus for this research. In the case of the
27
+ U.S.S. Stark, the commander made the decision to
28
+ not engage an inbound aircraft which was be-
29
+ lieved to not be a threat to his ship, and 27 U.S.
30
+ naval personnel lost their lives as a result. In the
31
+ case of the U.S.S. Vincennes, the commander
32
+ made the decision to engage the inbound aircraft
33
+ believing it was a threat to his ship—which turned
34
+ out to be a commercial airliner—and all personnel
35
+ aboard the airliner were killed as a result. In rec-
36
+ ognition of the complex and difficult decisions
37
+ required in these types of situations the Tactical
38
+ Decision Making Under Stress (TADMUS) pro-
39
+ gram was initiated to conduct research in the areas
40
+ of human factors and training technology. The
41
+ objective is to develop and apply principles that
42
+ can help avoid these types of situations in the fu-
43
+ ture. This paper, and a two companion papers
44
+ (Hutchins, Kelly, & Morrison, 1996; Kelly,
45
+ Hutchins, & Morrison, 1996), report on a multi-year, multi-experiment research effort conductedunder the TADMUS program to apply recent de-
46
+ velopments in decision theory and human-system
47
+ interaction technology to the design of a decision
48
+ support system (DSS) for enhancing tactical deci-
49
+ sion making under highly complex conditions.1.1 "Naturalistic" and Classical Decision-Making Para-digmsIn the same time frame that these tragic acci-dents occurred, a radical shift was occurring in the
50
+ way psychologists viewed human decision mak-
51
+ ing. Research was now focused on experienced
52
+ decision makers performing their normal tasks in
53
+ natural settings. "Naturalistic" decision-making
54
+ research studies the decision strategies people ac-
55
+ tually use in bringing their expertise to bear under
56
+ challenging real-world conditions. Decision
57
+ making milieus encompassed under the naturalis-
58
+ tic paradigm include hospital emergency rooms,
59
+ aircrew flight coordination, military command and
60
+ control settings, process control systems, and po-
61
+ lice and fire units.The naturalistic perspective, also known as"everyday cognition," is based on a belief that
62
+ cognitive functions "elicited in natural settings
63
+ (are) likely to differ, either quantitatively, or
64
+ qualitatively, from those that occur in artificial or
65
+ contrived situations, and results from sterile and
66
+ contrived situations may not generalize to less
67
+ constrained and more natural environments"
68
+ (Salthouse, 1992, p. 982). There is a growing
69
+ body of work that demonstrates that experienced,
70
+ real-world decision makers rarely use traditional
71
+ resource intensive strategies to make decisions in
72
+ the face of dynamic, adverse conditions and time-
73
+ pressure (Kaempf & Militelo, 1992; Klein, 1989;
74
+ 1993). Instead, experts rely on their abilities to
75
+ recognize and appropriately classify situations:
76
+ these abilities are based on having much experi-
77
+ ence in the task domain. Once they know what
78
+ they are facing they also tend to know what re-
79
+ sponse option to apply, based on retrieval from
80
+ memory of typical responses and outcomes that
81
+ worked well in past similar situations. They use
82
+ the limited time available to evaluate the feasibil-
83
+ ity of that option before implementing it. Experi-
84
+ enced decision makers recognize the situation or
85
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 3Monterey, Ca.scenario based on a comparison of the features ofthe current situation with stored memory represen-
86
+ tations, or schemata. Schemata are highly inter-
87
+ con-nected clusters of knowledge concerning cer-
88
+ tain situations, or particular problem types, and
89
+ associated actions or solution procedures (Fede-
90
+ rico, 1995). Once the situation is recognized, so-
91
+ lutions are stimulated by activation of these mem-
92
+ ory representations.In contrast to the naturalistic perspective, ear-lier analytical methods applied in decision support
93
+ systems were primarily used for option generation
94
+ and evaluation, rather than for situation assess-
95
+ ment. Traditional decision theorists argue that op-
96
+ timum decision making involves thorough analy-
97
+ sis of all the available data and the evaluation of
98
+ all possible hypotheses; these approaches tend to
99
+ rely on extensive calculations designed to arrive at
100
+ optimal solutions. Processes for making decisions
101
+ where the person weighs the pros and cons of
102
+ various options and selects the one option that
103
+ provides the most benefit are described in the de-
104
+ cision making literature. The extensive time re-
105
+ quirements and complicated mathematical calcu-
106
+ lations involved (e.g., Multi-Attribute Utility
107
+ Analysis), however, make these approaches unre-
108
+ alistic for situations requiring rapid decision
109
+ making. Generally, these analytical strategies in-
110
+ volve the following steps (Kaempf & Militello,
111
+ 1992): specify all relevant features of the task;
112
+  identify the full range of options;
113
+  identify the key evaluation dimensions;
114
+  identify weights for each dimension;
115
+  rate each option on each dimension;
116
+  tabulate the results, and
117
+  select the best option.These analytical strategies may be appropriate for
118
+ inexperienced subjects making decisions about
119
+ novel tasks, but not for experienced personnel
120
+ making real-time decisions. In natural settings,
121
+ time constraints and the difficulty in assigning
122
+ weights and rating the dimensions involved render
123
+ classical analysis techniques untenable. Typically
124
+ in realistic settings experts employ recognition-
125
+ based reasoning, not classical analytical ap-
126
+ proaches. Experienced decision makers use their
127
+ extensive knowledge to seek information, identifyand interpret the problem, understand the signifi-cance, derive the intention (where possible),
128
+ model the situation (as time allows), select the
129
+ action, evaluate the choice, and anticipate the con-
130
+ sequence. "This decision cycle is distinctly differ-
131
+ ent from classical models, which are based on the
132
+ assumption that all options, outcomes, and prefer-
133
+ ences are known and calculated in advance"
134
+ (Federico, 1995, p. 106).Traditionally, research in decision making hasbeen directed largely toward situations in which
135
+ (1) decision makers have sufficient time to gener-
136
+ ate options, conduct option assessment, and select
137
+ a course of action; (2) the consequences of an in-
138
+ correct response are not immediately severe; (3)
139
+ decisions are reached via consensus of a group;
140
+ and (4) workload is manageable. Little research
141
+ has been conducted into the development of tacti-
142
+ cal decision support systems for use in naturalistic
143
+ situations characterized by time pressure, high
144
+ risk, uncertainty and information ambiguity, high
145
+ workload, team coordination demands and task
146
+ complexity (ONT, 1992).The central hypothesis for the research re-ported here is that presenting decision makers
147
+ with decision support tools which were designed
148
+ to parallel the cognitive strategies employed by
149
+ experts, as observed in naturalistic settings, will
150
+ reduce the number of decision making errors. This
151
+ is accomplished by developing the architecture
152
+ and algorithms to process information the same
153
+ way research indicates humans do under similar
154
+ circumstances.2 Tactical Decision Making TasksThe global tactical decision-making task involvesidentification of and responding to numerous
155
+ contacts. When an aircraft (or a surface contact) is
156
+ detected the CIC personnel work as a team to de-
157
+ termine the identity and to try to determine
158
+ whether or not the aircraft poses a threat. The high
159
+ degree of inherent ambiguity associated with
160
+ contact information can often make threat assess-
161
+ ment a very difficult task. This is because many
162
+ pieces of data fit multiple hypotheses regarding
163
+ threat assessment. The global response choices
164
+ (that is, engage, monitor, do nothing) are largely
165
+ determined by the ship's orders and the current
166
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 4Monterey, Ca.geopolitical situation. Specific actions (such as,change course, issue verbal warnings, illuminate
167
+ with radar, challenge with other sensors, etc.) de-
168
+ pend on the local conditions and the relative posi-
169
+ tions of the inbound contact of interest and own-
170
+ ship. Determining which of these actions is likely
171
+ to be effective depends on maintaining an accurate
172
+ threat assessment which requires "continual up-
173
+ dating in accordance with recurrent situation as-
174
+ sessments" (Sarter & Woods, 1991, p. 52).This decision problem presents a highly chal-lenging cognitive task, that is, making inferences
175
+ and deductions from incomplete and uncertain
176
+ information derived from multiple sources and
177
+ relating to several concurrent threats (or potential
178
+ threats) under time-compressed conditions. The
179
+ cognitive functions performed by the tactical deci-
180
+ sion maker are both data and resource limited
181
+ (Norman & Bobrow, 1975). Decisions are re-
182
+ source limited by the mental resources of the deci-
183
+ sion makers, who must maintain large amounts of
184
+ information in memory under conditions of high
185
+ workload and stress. The decisions are data lim-
186
+ ited by the inability of the sensors to provide
187
+ complete, error-free, unambiguous data to support
188
+ the identification process. In particular, the ex-
189
+ perimental scenarios were designed to follow the
190
+ pattern of being set in an ambiguous situation
191
+ where one or more threats of uncertain origin and
192
+ uncertain intent approach either own ship or the
193
+ ship being protected and may not respond to
194
+ warnings. Scenarios were designed to be highly
195
+ ambiguous, as this quality of uncertainty is in-
196
+ dicative of the types of decisions to be made in
197
+ current and future scenarios.2.1 Threat AssessmentIn the antiair warfare problem, threat assess-ment is particularly difficult because the available
198
+ information is often incomplete or ambiguous.
199
+ The ambiguity could be due to (a) the information
200
+ transmission characteristics of the transmission
201
+ medium, such as a radar transmission or a radio
202
+ report that is only intercepted on an intermittent
203
+ basis, (b) deliberate deceptive actions (such as ra-
204
+ dar jamming) by the pilot flying the aircraft, or (c)
205
+ the overlapping classification categories typical of
206
+ many parameter measure-ments. For example, air-craft can typically fly at altitudes ranging between2,000 and 40,000 feet. Generally, an aircraft that
207
+ is flying above 20,000 feet is not considered to be
208
+ a threat. Conversely, an aircraft below 10,000 ft is
209
+ considered to be more of a potential threat. How-
210
+ ever, aircraft flying in the middle range, (that is,
211
+ below 20,000 ft. and above 10,000 ft.) can be
212
+ much more difficult to categorize. Because many
213
+ aircraft do fly in this middle range, other variables
214
+ need to be considered in conjunction with altitude.
215
+ This same situation of overlapping categorization
216
+ categories exists for several other variables. These
217
+ variables include radars that are found on both
218
+ threat and non-threat platforms, country of origin,
219
+ and measures of course and speed. In the case of
220
+ speed, for example, when an aircraft flying at a
221
+ low altitude decreases speed this could be viewed
222
+ as indicative of a threat action (that is, slowing
223
+ down in order to obtain better targeting informa-
224
+ tion); however, at the same time, there could be
225
+ other viable explanations for an aircraft's de-
226
+ creasing speed.If the decision maker had access to all dataabout a contact approximately twelve variables
227
+ would be used to determine identity and to infer
228
+ intent. Two or three of these items, alone, do not
229
+ provide definitive answers because, in many
230
+ cases, these parameter values do not fall within
231
+ clear-cut ranges for a particular assessment cate-
232
+ gory (i.e., threat, non-threat). Thus, a single time
233
+ slice of information provides an incomplete pic-
234
+ ture of the situation. In the dynamic, ambiguous
235
+ conditions characteristic of littoral operations, the
236
+ rate and direction of change (data history) can
237
+ help one better assess the threat and predict the
238
+ future state of the situation. When the incoming
239
+ information changes over time, the integration of
240
+ information as it changes can help the user extract
241
+ the message (Kirshenbaum, 1992). The DSS was
242
+ designed to do precisely this: to facilitate the inte-
243
+ gration process and present a synthesized picture
244
+ of the situation to the user in a format that can be
245
+ quickly assimilated. The variables, that are used to
246
+ develop a threat assessment, can be divided into
247
+ two classes: sensor information (raw or computer-
248
+ processed information) and the contact's response,
249
+ or lack of response, to actions taken by the team.
250
+ These other actions, and the integration of the in-
251
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 5Monterey, Ca.formation received via the contact's response orlack of response to them, are necessary to clarify
252
+ the tactical picture.2.2 Situation AwarenessWhile recent increasing interest among re-searchers regarding the concept of situation
253
+ awareness (SA) has generated a debate on the pre-
254
+ cise definition of this term most researchers ac-
255
+ knowledge the importance of the concept. In gen-
256
+ eral, SA refers to the decision maker's moment-
257
+ by-moment ability to monitor and understand the
258
+ state of the complex system and its environment
259
+ (Adams, Tenney, & Pew, 1995). These authors
260
+ state the essential idea which is that when emer-
261
+ gencies arise, the completeness and accuracy of
262
+ the decision maker's SA are critical to the ability
263
+ to make decisions, revise plans, and manage the
264
+ system. Specific decision-making tasks included
265
+ under SA include the ability to: (1) maintain an
266
+ accurate perception of the surrounding environ-
267
+ ment (both internal and external to the ship); (2)
268
+ identify problems and/or potential problems; (3)
269
+ recognize a need for action; (4) note deviations in
270
+ the mission; and (5) maintain awareness of tasks
271
+ performed (Shrestha, Prince, Baker and Salas,
272
+ 1995). To maintain an accurate SA the decision
273
+ maker should take into account both information
274
+ that is available and that which can be activated
275
+ from memory (Sarter and Woods, 1991).However, a difficulty arises as a result of theheavy workload imposed by this process. When
276
+ the decision maker is faced with several concur-
277
+ rent contacts of interest, all of which have numer-
278
+ ous associated data items (i.e., as many as a
279
+ dozen), some or all of which may change over the
280
+ course of the scenario (e.g., intelligence, active
281
+ radar emitters, various kinematic parameters, etc.)
282
+ memory load easily exceeds human capacity.
283
+ Moreover, changing parameters may impart dif-
284
+ ferent interpretations to what is occurring. In some
285
+ cases the moment-to-moment attentional demands
286
+ of a tactical situation are relentless and unforgiv-
287
+ ing (such as, a terrorist aircraft directly inbound
288
+ toward "own-ship" which can result in "task fixa-
289
+ tion"), sometimes relevant background knowledge
290
+ is unavoidably incomplete (such as, an unfamiliar
291
+ aircraft), and sometimes the decision maker is al-ready thinking and working as hard as possible,even when there are no unanticipated events when
292
+ there is a high contact density (Adams, et al,
293
+ 1995). These instances provide a few illustrations
294
+ of situations that can degrade situation awareness.Complex information gathering and process-ing systems have been designed to aid the deci-
295
+ sion-maker in the past. However, these systems
296
+ often increase the decision-maker’s burden due to
297
+ the inherent system complexity and the failure to
298
+ design them in a way that they will fit the user's
299
+ cognitive processing limitations. Often, these
300
+ systems require operators to perform difficult
301
+ cognitive tasks under heavy workloads. They must
302
+ perceive, synthesize and determine the relevance
303
+ of a continual stream of incoming information,
304
+ often pertaining to several concurrent contacts,
305
+ while projecting future anticipated events and
306
+ making decisions regarding actions to be taken.
307
+ Decision makers must assess, compare, and re-
308
+ solve conflicting information, while making diffi-
309
+ cult judgments, and remembering the status of
310
+ critical contacts along with the contact's response
311
+ to actions taken by the CIC team. These decision-
312
+ making tasks are interleaved with other required
313
+ tasks, such as keeping other team members in-
314
+ formed (both on and off the ship). Furthermore,
315
+ these complex tasks are performed under condi-
316
+ tions where adverse environmental (noise, vibra-
317
+ tion, temperature extremes, etc.) and internal
318
+ stressors (boredom, fatigue, anxiety, and fear) are
319
+ part of the environment.3 Decision Support PrinciplesA case has been made that previous generations ofdecision support systems, which focus primarily
320
+ on solution optimization and base decision sup-
321
+ port on normative models of human decision
322
+ making, are less applicable than a DSS that par-
323
+ allels the cognitive strategies used by domain ex-
324
+ perts in situations characterized by time-
325
+ constrained situations with uncertain and ambigu-
326
+ ous data (Smith & Grossman, 1993). These
327
+ authors point out that rarely, if ever, were earlier
328
+ tactical decision aids intended as psychological
329
+ models of human cognitive behavior. Instead,
330
+ these aids performed "complex and burdensome
331
+ calculations, reducing the work-load on personnel,
332
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 6Monterey, Ca.speeding up the dissemination of information, andproviding more time for command decision mak-
333
+ ing" (Tolcott, 1991, p. 44).3.1 Feature Matching and Story GenerationWe have applied two of these new models ofhuman decision making—which parallel the cog-
334
+ nitive strategies used by domain experts—to the
335
+ design of a DSS for enhancing antiair warfare
336
+ tactical decision making. These two models that
337
+ people use in assessing a situation are feature
338
+ matching and story generation. The feature
339
+ matching model, described by Noble (1989), in-
340
+ volves an organization of memory, or "schemas,"
341
+ and information-processing where decision mak-
342
+ ers use their previous experiences to assess a
343
+ situation and identify promising actions. Incoming
344
+ information is categorized, selected, edited, and
345
+ organized on the basis of a person's general
346
+ knowledge about a domain. Both story generation
347
+ and feature matching occur under conditions
348
+ where a large base of implication-rich, condition-
349
+ ally dependent pieces of evidence must be evalu-
350
+ ated before choosing an alternative from a set of
351
+ prospective courses of action. The feature match-
352
+ ing model applies a spatio-temporal dependence,
353
+ whereas story generation is an example of causal
354
+ dependence. According to the explanation-based
355
+ model, decision makers construct a causal model
356
+ to explain the available evidence (Pennington &
357
+ Hastie, 1993). At the same time, the decision
358
+ maker creates a set of alternatives from which an
359
+ action will be chosen. A decision is made when a
360
+ story is successfully matched to an alternative in
361
+ the choice set. Story generation occurs in complex
362
+ situations where the decision maker may not have
363
+ all the necessary information or when a series of
364
+ facts may appear to contradict each other. The de-
365
+ cision maker must then develop causal links be-
366
+ tween these facts to produce a coherent picture of
367
+ the situation (Klein, 1989; 1993).The explanation-based reasoning model isbased on research which found that jurors develop
368
+ a narrative story to organize trial information
369
+ where causal and intentional relations between
370
+ events are central (Pennington & Hastie, 1992).
371
+ Pennington & Hastie propose four certainty prin-
372
+ ciples—coverage, coherence, uniqueness, andgoodness-of-fit—that govern (i) which story willbe accepted, (ii) which decision will be selected,and the (iii) confidence or degree of certainty withwhich a particular decision will be made. This or-
373
+ ganization of the evidence by the decision maker
374
+ is believed to facilitate evidence comprehension.
375
+ A central component of this model is that the
376
+ story the juror constructs determines the juror's
377
+ decision. This explanation-based decision process
378
+ is employed when the body of evidence relevant
379
+ to a decision is large, complex, and the implica-
380
+ tions of its components are interdependent.Feature matching, also referred to as the rec-ognition-primed decision (RPD) model, "occurs
381
+ when the decision maker recognizes the features
382
+ of the present situation as similar or identical to
383
+ those of a previous situation" (Kaempf & Militelo,
384
+ 1992, p. 6). An adequate match triggers recall of
385
+ information learned about this type of situation:
386
+ (a) plausible goals, (b) critical cues to be moni-
387
+ tored, (c) expectations of what should happen, and
388
+ (d) a course of action that worked in similar situa-
389
+ tions. According to this recent approach, expert
390
+ decision makers may rely on well-developed
391
+ memory representations to guide decision making
392
+ in new (but similar) situations. The RPD model of
393
+ decision making fuses two processes—situation
394
+ assessment and mental simulation (Klein, 1993).
395
+ In the simplest case the situation is recognized as
396
+ familiar or prototypical, using feature matching,
397
+ and the obvious response is implemented. In a
398
+ more complex case the decision maker performs a
399
+ conscious evaluation of the response, using men-
400
+ tal simulation to uncover problems prior to im-
401
+ plementing the response. In the most complex
402
+ case the evaluation reveals flaws requiring modi-
403
+ fication, or the option is judged inadequate and
404
+ rejected in favor of the next most typical reaction.3.2 Situation AssessmentIn general, the overall task of responding toantiair warfare scenarios consists of situation as-
405
+ sessment ("what's going on") and course of action
406
+ selection ("what to do about it"). Recent theories
407
+ of decision making emphasize the importance of
408
+ situation assessment for good decision making in
409
+ naturalistic, event-driven situations. Moreover,
410
+ they stress that decisions regarding actions to be
411
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 7Monterey, Ca.taken are a by-product of developing the situationawareness that precedes action selection. Klein
412
+ (1989) has found that usually the situation itself
413
+ either determines or constrains the response op-
414
+ tions and that experienced decision makers make
415
+ up to 90% of all decisions without considering
416
+ alternatives. If the situation appears similar to one
417
+ that the decision maker has previously experi-
418
+ enced, the pattern will be recognized and the
419
+ course of action is usually immediately obvious.
420
+ On the other hand, if the situation does not seem
421
+ familiar complex RPD will be involved where the
422
+ decision maker adjusts the option after evaluating
423
+ it.Additional evidence was found in the specifictask domain of interest to the TADMUS program
424
+ which added support to these findings on the way
425
+ real-world decision makers make decisions in the
426
+ context of their normal jobs. Research was con-
427
+ ducted to determine decision requirements for
428
+ command-level decision makers in the combat
429
+ information center (CIC) of an Aegis cruiser.
430
+ Analysis of 14 incidents from actual problems re-
431
+ vealed 183 decisions. Of these, 103 concerned
432
+ situation assessments. Results obtained after ana-
433
+ lysts coded these situation assessments indicated
434
+ that decision makers arrived at approximately
435
+ 87% of their situation assessments through feature
436
+ matching and the remaining 13% through story
437
+ generation (Kaempf, Wolf, & Miller, 1993). The
438
+ other eighty decisions that were identified, from
439
+ analysis of the real-world incidents mentioned
440
+ above, involved course of action selection. These
441
+ course of action decisions served a variety of
442
+ functions, although, relatively few were intended
443
+ to end the incident. Twenty were intended as a
444
+ final course of action decision; 14 were imple-
445
+ mented to obtain more information, 22 to manage
446
+ resources, and 24 to put themselves in a more fa-
447
+ vorable tactical position. A recognition-based
448
+ strategy was also used by decision makers to de-
449
+ velop a course of action, accounting for 95% of
450
+ the actions taken in the 14 incidents. The decision
451
+ makers generated and compared multiple options
452
+ in only 5% of the cases. In line with these find-
453
+ ings, the TADMUS program has adopted the po-
454
+ sition that decision aiding systems should assist in
455
+ the decision making process, and focus on aidingthe situation assessment portion of the decision-making task.A DSS was developed to support decision-making processes which research has shown are
456
+ used by decision makers in real-world settings
457
+ (Hutchins, Kelly, & Morrison, 1996). Specifi-
458
+ cally, the DSS parallels the strategies used by ex-
459
+ perienced decision makers to perform situation
460
+ assessment (Nobel, 1989; 1993). This approach to
461
+ supporting the user's intuitive approach to dealing
462
+ with dynamic decision-making situations should
463
+ produce tools that are both more easily understood
464
+ and used, and that more effectively "exploit the
465
+ decision maker's knowledge and expertise that
466
+ might facilitate adaptation to complex, novel
467
+ situations" (Cohen, 1993, p. 265).4 Human-System Interaction PrinciplesThe vast majority of research on human-computerinteraction design has been devoted to character-
468
+ istics of displays that impact human perception,
469
+ such as symbol legibility or detectability, and on
470
+ relatively simple cognitive functions such as
471
+ memory tasks. Fewer efforts have been devoted to
472
+ understanding the effects of the format and man-
473
+ ner in which information is presented on more
474
+ complex levels of human cognition such as deci-
475
+ sion making. Consequently, principles that can be
476
+ applied to the design of the interface between the
477
+ user and a decision support system for the purpose
478
+ of enhancing cognitive changing situations, are
479
+ not available to any significant degree (ONT,
480
+ 1992).4.1 Graphic PresentationsSeveral advantages are offered by graphicpresentations over a text-based presentation for-
481
+ mat (Larkin and Simon, 1987). Graphic presenta-
482
+ tions should (1) reduce the amount of mental
483
+ computation required to perform tasks; and (2)
484
+ allow users to spend less time searching for
485
+ needed information. Casner (1991) elaborated on
486
+ these ideas and found that graphics allow users to
487
+ substitute less demanding perceptual opera-tions
488
+ for more complex logical operations. For exam-
489
+ ple, determining a change in altitude (and the de-
490
+ gree of change) is immediately apparent when the
491
+ user glances at the track history module. (Note
492
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 8Monterey, Ca.that the words contact and track can be used inter-changeably. The reader is referred to Figure 1.)The objective for the track history module isto facilitate the contact identification process by
493
+ providing information that is integrated in a way
494
+ that supports a recognitional decision strategy.
495
+ This module depicts a contact’s speed, altitude,
496
+ course and range on a two-dimensional graphical
497
+ display along with a geometric representation of
498
+ both the contact’s weapon release envelope and
499
+ own-ship’s weapons coverage. A large amount of
500
+ parametric data is portrayed graphically for rapidassimilation by the user. The user can see, at aglance, a synthesized picture of the contact’s be-
501
+ havior. Compare this rather simple perceptual op-
502
+ eration with the more complex logical operation
503
+ involved in current operational systems which re-
504
+ quire the user to recall and subtract numerical val-
505
+ ues for past and current altitudes.Graphics also allow users to omit steps thatare otherwise necessary when a task is performed
506
+ without a graphic. An example of this advantage
507
+ isFigure 1. Decision Support System Display Modules.also illustrated in the track history module whichincludes templates indicating weapon's coverage
508
+ for both the inbound contact and "own-ship." To
509
+ determine whether the aircraft is within its
510
+ weapon's launch range there is no need to recall
511
+ the specific launch range values and then compare
512
+ them with the aircraft's current range. Instead, the
513
+ user can determine if the aircraft is within its
514
+ launch range by a quick glance at the display.Graphics help users save time when searchingfor needed information when several related di-
515
+ mensions of information are encoded in a single
516
+ graphical object. This is accomplished by inte-
517
+ grating the kinematic parameters of speed, course,
518
+ altitude, bearing, and range for a contact. The user
519
+ can see, at a glance, a synthesized picture of the
520
+ contact's behavior. Compare this process with
521
+ reading, in a text-based format, the individual pa-
522
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 9Monterey, Ca.rameters which need to be integrated by the userinto a coherent picture of the contact's behavior.5 Limited Cognitive Processing CapabilitiesSince there are limits to the cognitive processingcapability of humans, it is important for the sys-
523
+ tem to provide the needed information in a format
524
+ that best supports the user operating under dy-
525
+ namic decision-making conditions. It may be the
526
+ case that current systems are inadequate to support
527
+ the cognitive processing demands required by
528
+ certain littoral scenarios. For example, according
529
+ to Gruner (1990, p. 41), the U.S.S. Vincennes of-
530
+ ficers and system operators "could not make better
531
+ decisions because they did not have time to con-
532
+ firm or deny the information uncertainties pre-
533
+ sented them." Gruner maintains that the rapid pace
534
+ involved in these types of situations can exceed
535
+ the capacity of the human to comprehend the
536
+ rapid flow of information presented by complex
537
+ systems. In the case of the Vincennes, the CIC
538
+ team had three minutes and 40 seconds to make
539
+ their decision. This includes the time required for
540
+ the operators to perceive and interpret sensor data
541
+ and for the commanding officer to make informed
542
+ judgments from these data (Roberts & Dotterway,
543
+ 1995). The result of the human's limited cognitive
544
+ processing capabilities is that the decision makers
545
+ may fail to remember critical pieces of data,
546
+ overlook stored information, draw hasty conclu-
547
+ sions, and produce flawed answers.Evidence of the effects of limitations in mem-ory and shared attention capacity on human deci-
548
+ sion making were found during baseline testing
549
+ (Hutchins & Kowalski, 1994; Hutchins & Westra,
550
+ 1995) and during empirical evaluation of the DSS
551
+ (Kelly, Hutchins, & Morrison, 1996; Kelly, Mor-
552
+ rison, & Hutchins. 1996).Simon (1978, p. 273) states, "...the human in-formation processing system...operates almost en-
553
+ tirely serially, one process at a time, rather than in
554
+ parallel fashion. This seriality is reflected in the
555
+ narrowness of its momentary focus of attention."
556
+ However, the AAW problem forces the decision
557
+ maker to operate in a parallel processing mode
558
+ when several contacts demand attention at the
559
+ same time. The requirement to monitor and
560
+ maintain an accurate SA for these concurrentcontacts, over the course of the evolving situation,imposes an additional load of strategically man-
561
+ aging the overall situation. Several researchers
562
+ have argued that "managing the attentional and
563
+ conceptual processes that permit cogent SA in-
564
+ volves significant cognitive resources" (Adams, et
565
+ al, 1995, p. 91; Endsley, 1988). The tasks of pri-
566
+ oritizing contacts and the associated actions to be
567
+ taken by the team, updating the status of critical
568
+ contacts, responding to the other requisite tasks in
569
+ the queue and, more generally, of strategically
570
+ managing the workload of current multitask sys-
571
+ tems under dynamically changing scenarios can
572
+ place an unrealistic cognitive load on the decision
573
+ maker.A major advantage offered by the experimen-tal DSS is that it should "buy time" for the user by
574
+ (1) performing many of the cognitive processing
575
+ tasks for the user and (2) by presenting informa-
576
+ tion in graphic format. The DSS will synthesize
577
+ much of the information used to develop situation
578
+ awareness and present a coherent picture of the
579
+ situation to the user. This integrated picture will
580
+ be portrayed graphically—rather than in the cur-
581
+ rent text-based format—which should further re-
582
+ duce the amount of time required to assimilate
583
+ this information. By performing several informa-
584
+ tion processing steps for the decision maker the
585
+ decision maker's limited cognitive resources can
586
+ be used for the types of decisions which require
587
+ human abilities (e.g., the decision on whether to
588
+ engage).5.1 Working Memory RequirementsAn essential information processing step re-quired by this task—and one which levies a heavy
589
+ load on working memory—involves integrating
590
+ kinematic and sensor variables and maintaining an
591
+ awareness of changes in these variables over time.
592
+ Changes in a contact's behavior such as, decreas-
593
+ ing altitude, increasing speed, changes in elec-
594
+ tronic emissions, etc., can provide key indicators
595
+ of possible hostile intent. With current systems,
596
+ the decision maker receives numerous reports
597
+ from CIC team members who provide various
598
+ pieces of the overall tactical picture (such as,
599
+ kinematic parameter values, active electronic
600
+ emitter identifications, and behavioral responses
601
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 10Monterey, Ca.of the contact in response to queries by the team)regarding a particular contact. Some of this infor-
602
+ mation is also displayed in a text-based format for
603
+ the user when a contact is "hooked" (that is, se-
604
+ lected for display) by the decision maker. How-
605
+ ever, to recognize a change in certain variables,
606
+ current systems require the user to retain parame-
607
+ ter values in short-term memory in order to recog-
608
+ nize a change in the parameter, such as altitude.When the decision maker is monitoring sev-eral concurrent contacts (such as, cycling through
609
+ three or four contacts in a 1-minute period) human
610
+ working memory capabilities may quickly be sur-
611
+ passed. To detect a change in a critical parameter
612
+ value, the decision maker must maintain the pa-
613
+ rameter values for the contacts of interest in
614
+ working memory as he or she cycles between sev-
615
+ eral contacts. For example, the decision maker
616
+ must be able to recall that contact 7022 was at
617
+ 14,000 ft. altitude one minute ago, and then sub-
618
+ tract the current altitude value of 10,000 ft., which
619
+ will then indicate the aircraft is in a rapid descent.
620
+ The DSS was developed to aid the decision maker
621
+ by performing several of these cognitive process-
622
+ ing tasks, thus, reducing the cognitive load for the
623
+ user. By presenting the synthesized picture of the
624
+ contact's behavior over time, through the use of
625
+ graphical displays, critical changes should be
626
+ immediately apparent to the user.A second memory-intensive task involvesmaintaining, in working memory, a current list of
627
+ actions taken by team members, the contact's re-
628
+ sponse to these actions taken by the CIC team,
629
+ and pending actions. Research has established that
630
+ "memory is limited and that list maintenance is
631
+ effortful and fallible—more so if the list must be
632
+ ordered and still more if the membership of the
633
+ list must be dynamically reordered and modified
634
+ during retention" (Bower, 1970, as cited in Ad-
635
+ ams, et al, 1995, p. 91). The DSS should reduce
636
+ the cognitive effort required for distributing atten-
637
+ tion among the many contacts to be attended to
638
+ and actions that are required. Working memory
639
+ requirements should be reduced by having the
640
+ DSS act as an intelligent "assistant," reminding
641
+ the user regarding what actions are to be taken and
642
+ when the actions are to be taken.A third way the DSS will reduce memory andinformation processing requirements is by dis-
643
+ playing templates depicting weapons' envelopes
644
+ for both the inbound contact and "own-ship." This
645
+ should facilitate critical comparisons and judg-
646
+ ments regarding timing of actions. During a sce-
647
+ nario decision makers have to either rely on mem-
648
+ ory to recall the launch range for various weapons
649
+ or query a team member for this information. Both
650
+ of these methods waste limited resources. The
651
+ high workload and high tempo characteristic of
652
+ littoral scenarios produce a stress-ful decision-
653
+ making environment. The phenom-enon that in-
654
+ creasing stress leads to decreasing working mem-
655
+ ory is well documented (e.g., Hockey, 1986). The
656
+ latter method for obtaining the desired informa-
657
+ tion wastes limited resources by increasing the
658
+ communications load and requiring more time to
659
+ wait for a team members' response to the queryUnder these high-tempo and high workloadconditions human memory and attentional re-
660
+ sources can easily be surpassed. Several cogni-
661
+ tively resource intensive information processing
662
+ steps are eliminated for the human decision maker
663
+ by having them performed by the DSS. We pre-
664
+ dict that the decision support tools will reduce the
665
+ cognitive workload imposed on the decision
666
+ maker in the following three ways: (1) by reduc-
667
+ ing the amount of information processing to be
668
+ performed, (2) reducing working memory re-
669
+ quirements, and (3) assisting the user in allocating
670
+ limited attentional resources.5.2 Reducing Human ErrorThe study of human cognitive processes andrelated error mechanisms has gained rapidly in-
671
+ creasing interest in the past decade. Rasmussen
672
+ (1987) argues that the emphasis in attempting to
673
+ understand human errors must shift from tasks to
674
+ the human-task mismatch. For example, Gruner
675
+ (p. 39), in discussing the Vincennes incident,
676
+ maintains that "the system was poorly suited for
677
+ use by human beings during rapid military action."
678
+ He ascribes this lack of suitability to a human-
679
+ machine mismatch between the rate of data flow
680
+ possible with modern computer systems that can
681
+ process and display information at phenomenal
682
+ data rates and the "comprehension capability of
683
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 11Monterey, Ca.users which has remained almost static for thou-sands of years." This causal approach to under-
684
+ standing human error is based on the premise that
685
+ errors are rarely random and can be traced to
686
+ causes and contributing factors. Once these con-
687
+ tributing factors are identified they can be miti-
688
+ gated.The impact and vulnerability of systems andhuman interfaces, because of incompatibilities
689
+ between the way people perceive, think, and act,
690
+ are documented in the popular and technical lit-
691
+ erature (Buck, 1989; Casey, 1993; Norman, 1988;
692
+ Perrow, 1984; Wilson & Zarakas, 1978). Newly
693
+ developed systems will succeed or fail based on
694
+ our ability to minimize these incompat-ibilities
695
+ between the characteristics of the things we create
696
+ and the way we use them. There are many well-
697
+ documented instances of critical systems or pa-
698
+ rameter changes going unnoticed or unheeded be-
699
+ cause the operating procedures, or the human ma-
700
+ chine interface, provided no historical trace. For
701
+ example, an unnoticed increase in altitude con-
702
+ tributed to the shoot down of the Iranian airbus by
703
+ a U.S. Navy ship—when the team mistakenly be-
704
+ lieved the aircraft to be descending—because
705
+ there was no historical trace to make the aircraft's
706
+ actual increasing altitude apparent (Dotterway,1992). Five personnel in the U.S.S. Vincennes's
707
+ combat information center, all viewing separate
708
+ displays, reported the aircraft as descending while
709
+ the Aegis data tapes later revealed a flight pattern
710
+ of ascent (Roberts & Dotterway, 1995). One of the
711
+ official investiga-tions of this incident, the Fo-
712
+ garty Report (1988, p. 45), states that "stress, task
713
+ fixation, and an unconscious distortion of data
714
+ may have played a major role in this incident." A
715
+ panel of five psychologists from the American
716
+ Psychological Association who testified before
717
+ Congress concluded that there were "predictable
718
+ failings of human judgment under intense stress
719
+ compounded by complex technology [which]
720
+ clearly contributed to the accidental shooting of
721
+ Iranian airliner Flight 655" (APA, p. 4).It is generally accepted that between 60-80percent of the accidents and malfunctions in
722
+ transportation, manufacturing, process control,
723
+ weapon, and other systems are attributable to hu-
724
+ man error (Senders & Moray, 1991; Van Cott,1993; Weiner, 1994). Reducing tactical decisionmaking errors is one goal of the TADMUS pro-
725
+ gram. The following section presents a brief re-
726
+ view of an experiment conducted to develop a
727
+ baseline on tactical decision making performance
728
+ in response to fairly stressful scenarios. A com-
729
+ panion paper (Hutchins, Kelly, & Morrison, 1996)
730
+ describes the experimental DSS modules and the
731
+ way they are hypothesized to enhance tactical de-
732
+ cision making performance.6 TADMUS Baseline ExperimentEarly research involved data collection in the De-cision-Making Evaluation Facility for Tactical
733
+ Teams (DEFTT) Laboratory using simulated ex-
734
+ isting shipboard displays to establish a baseline on
735
+ decision-making performance. The purpose of this
736
+ effort was to document baseline decision-making
737
+ performance for experienced naval officers. Dur-
738
+ ing the baseline phase of testing, a detailed under-
739
+ standing was developed of the cognitive processes
740
+ underlying the various tasks involved in situation
741
+ assessment—and where the bottlenecks occur.
742
+ This understanding was then used to design the
743
+ way the information is presented to the user in or-
744
+ der to facilitate performance of the required tasks.6.1 SubjectsThis study focused on the command-level de-cision makers of an antiair warfare team on an
745
+ Aegis cruiser—the commanding officer and the
746
+ tactical action officer. Subjects in the study con-
747
+ sisted of six commanding officer/tactical action
748
+ officer teams drawn from twelve active duty Na-
749
+ val personnel; some were from training com-
750
+ mands while others were from operational com-
751
+ mands aboard ship or assigned to group staffs.6.2 ProcedureData were collected in the DEFTT Labora-tory, a six-station test-bed environment that
752
+ simulates console positions in a Navy Aegis
753
+ cruiser combat information center. (For a detailed
754
+ description of the DEFTT Laboratory see Hutch-
755
+ ins, 1996.) Four stations were filled by confeder-
756
+ ates (active duty Navy personnel) who play antiair
757
+ warfare support-team member roles. These roles
758
+ included the antiair warfare coordinator, identifi-
759
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 12Monterey, Ca.cation supervisor, tactical information coordina-tor, and electronic warfare supervisor. After ap-
760
+ proximately 1 1/2 hours of orientation to the labo-
761
+ ratory and training in the use of the computer con-
762
+ soles the subjects engaged in four scenarios. The
763
+ scenarios were each about 25 minutes in length
764
+ and contained between 11 and 14 contacts of in-
765
+ terest per scenario, in addition to numerous back-
766
+ ground contacts.6.3 Treatment of DataTeam communications were recorded on amultichannel audio recorder; these included all
767
+ intra-team exchanges, as well as all communica-
768
+ tions with simulated off-ship personnel. Audio
769
+ tapes were used to produce verbatim, time-
770
+ stamped transcripts of all team communications.
771
+ A modified version of the TapRoot® Incident In-
772
+ vestigation System (Paradies, 1991; Paradies and
773
+ Unger, 1991) was then applied to identify errors.
774
+ The objective was to identify tactically significant
775
+ errors committed during the scenario. Tactically
776
+ significant errors were defined as those errors that
777
+ may lead to loss of life or significant political em-
778
+ barrassment. The following criteria were used for
779
+ counting an error as tactically significant: (1) loss
780
+ of situation awareness, (2) failure to take defen-
781
+ sive action when within the weapon's range of an
782
+ approaching contact, or (3) a violation of rules of
783
+ engagement (ROE). Video recordings were made
784
+ of the commanding officer and tactical action of-
785
+ ficer computer screens. Detailed analyses of all
786
+ audio and video recordings were conducted. (For
787
+ a more detailed coverage of the methodology and
788
+ results see Hutchins and Westra, in preparation).6.4 ResultsThe complex, time-constrained, decision-making situations embodied in the experimental
789
+ scenarios resulted in a large number of decision
790
+ errors. The mean number of tactically significant
791
+ errors documented across six teams and four sce-
792
+ narios was 14; the number of errors ranged from
793
+ nine to twenty-two. The standard deviation was
794
+ 3.7. Subjects performed an average of 50% of the
795
+ required behaviors as specified in the rules of en-
796
+ gagement. The ordinal agreement between three
797
+ raters (navy subject matter experts) on error countranks from TapRoot® analyses was computed.Results showed a high degree of agreement with
798
+ the Kendall's W of .93 indicating that 93% of the
799
+ possible rank variance is accounted for.6.4.1 Decision-Making ErrorsDetailed examinations of the informationprocessing sequences performed during tactical
800
+ decision making have revealed a variety of errors.
801
+ On average, subjects failed to take required ac-
802
+ tions, about half of the time. Explanations based
803
+ in the cognitive psychology literature have been
804
+ pursued, as a major goal of the TADMUS pro-
805
+ gram is to develop a DSS based on an under-
806
+ standing of the way in which human decision
807
+ makers actually process information under rapidly
808
+ evolving situations.The majority of documented errors involvederrors of omission, that is, "failure to take defen-
809
+ sive measures” and "failure to adhere to ROE.”
810
+ Failure to take defensive measures included fail-
811
+ ure to take actions to defend own-ship when an
812
+ approaching aircraft had reached its weapon's re-
813
+ lease range. An example involved a case where
814
+ two contacts were within the specified ROE limit,
815
+ yet no action had been taken. The types of actions
816
+ included in the “failure to adhere to ROE” cate-
817
+ gory include failure to take action regarding the
818
+ items listed and defined below: (a) issuing warn-ings is part of the usual identification process andinvolves three levels of warnings with increasing
819
+ levels of urgency; (b) establish friendly force cri-teria refers to establishing a plan with otherfriendly ships in the area to coordinate how they
820
+ will respond to potential threats; (c) changes inkinematics/ identification friend or foe—subjectsare expected to notice significant kinematic
821
+ changes and/or identification friend or foe pa-
822
+ rameter changes; and (d) other identification pro-cedures includes actions such as illuminating withfire control radar.The other major category of tactically signifi-cant error involved “loss of SA.” Loss of SA er-
823
+ rors were grouped under errors of commission and
824
+ errors of omission and then further categorized
825
+ into subgroups. Fifty-five percent of the loss of
826
+ SA errors involved taking the wrong action (error
827
+ of commission) while 45% of the errors involved
828
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 13Monterey, Ca.failing to take some required action (error ofomission). Error categories included under errors
829
+ of commission involved incorrectly engaging a
830
+ track (3%), incorrectly warning a track (29%),
831
+ other incorrect actions (16%), and incorrect re-
832
+ porting (7%). The two instances of incorrectly en-
833
+ gaging an aircraft, which were F-1 Mirage aircraft,
834
+ were considered errors because the decision
835
+ maker failed to take certain actions prior to en-
836
+ gaging—not necessarily because the aircraft
837
+ should not have been engaged. The actions that
838
+ the decision makers failed to take involved ascer-
839
+ tain-ing the identification of the aircraft for one
840
+ case and failure to warn and illuminate the aircraft
841
+ prior to engaging for the second case. Most in-
842
+ stances of incorrectly issuing warnings to the air-
843
+ craft involved issuing the warning when the air-
844
+ craft was within its territorial airspace (that is, in-
845
+ side the 12 nautical mile limit which is interna-
846
+ tionally recognized as under control of that nation)
847
+ or issuing a warning at a level different from what
848
+ was required. Other incorrect actions included il-
849
+ luminating the aircraft, “locking up” with radar, or
850
+ ordering the aircraft to divert when the aircraft
851
+ was still within its territorial airspace. Incorrect
852
+ reporting involved inaccurate reports on the status
853
+ of the tactical situation (such as, indicating to the
854
+ battle group commander that certain actions had
855
+ been taken when they had not, misidentification of
856
+ an aircraft, or omitting critical tracks from a re-
857
+ port).Errors of omission categorized under the “lossof SA” category included: (a) failure to identify or
858
+ attend to a contact; (b) failure to take action (e.g.,
859
+ to issue “hold-fire” when a contact turned out-
860
+ bound); (c) failure to recognize a threat (e.g.,
861
+ designating an aircraft as a non-threat because it
862
+ had passed its closet point-of-approach, yet it was
863
+ still within missile-launch range); (d) instances of
864
+ confusion or forgetting (e.g., forgetting or ignor-
865
+ ing critical data, forgetting whether or not it had
866
+ been warned, illuminated, or “locked-on,” or for-
867
+ getting the aircraft's response, or lack of response
868
+ to these actions, forgetting the status of a contact,
869
+ and confusing contacts); (e) misperception of data
870
+ (e.g., reporting a contact as turning outbound
871
+ when it is still inbound); (f) unclear communica-
872
+ tion (issuing vague orders regarding actions to betaken by team member, such as, failure to specifywhich weapon system is to be used or which con-
873
+ tact is to be engaged).6.4.2 Cognitive explanationsThe cause of failures to take required actionsis, in many cases, attributed to the extremely high
874
+ task demands levied on the decision maker by the
875
+ scenario and the human decision-maker's limited
876
+ attentional resources. Many cases are also attrib-
877
+ uted to working memory limitations. Maintaining
878
+ an awareness of the status of each contact and the
879
+ status of many actions to be taken by the antiair
880
+ warfare team—which actions have been taken and
881
+ what the contact's response to the action was—
882
+ severely taxes the decision maker's working
883
+ memory. The high workload entailed in the
884
+ scenarios produces a highly time-compressed
885
+ decision making situation. This time-compressed
886
+ decision making situation—where attentional
887
+ resources and working memory capacity are
888
+ limited—do not allow the decision maker to
889
+ maintain accurate SA for all tracks at any given
890
+ time. We anticipate that the decision support
891
+ modules in the DSS will mitigate these types of
892
+ errors.Human information processing capabilities arenot well suited to dealing with a "multiplicity of
893
+ simultaneous and disjointed tasks. Thoughtful at-
894
+ tention is modular: People can consciously think
895
+ about only one thing at a time" (Adams, et al,
896
+ 1995, p. 92). As a result, they do not handle inter-
897
+ ruptions very well. Research indicates that when
898
+ an operator is faced with as few as two tasks that
899
+ consist of merely the detection or recognition of
900
+ simple signals, a cost may be incurred in terms of
901
+ a significant loss in sensitivity or time that can be
902
+ allocated to either by the requirement to divide or
903
+ switch attention between them (Broadbent, 1957;
904
+ Schneider and Detweiler, 1988; Swets, 1984).The memory demands of managing complex,multi-task situations can easily surpass human
905
+ limitations. The decision maker must not forget
906
+ any of the contacts or tasks requiring action. In
907
+ addition to remembering all the tasks needing at-
908
+ tention, however, are the complexities entailed in
909
+ keeping track of the data and substeps associated
910
+ with each contact and prior action. The aviation
911
+ literature provides many examples of incidents
912
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 14Monterey, Ca.with explanations similar to the root causes forerrors that were found in the TADMUS program.
913
+ One category includes the potentially disastrous
914
+ effects of interruptions in the task for air traffic
915
+ controllers and pilots. Similarly, in the AAW en-
916
+ vironment, momentary intervening attention to
917
+ another task or contact, or an interruption in a
918
+ procedure can leave the procedure, or processing
919
+ of a contact incomplete with potentially cata-
920
+ strophic results.A fairly consistent pattern of tactical decision-making errors was documented from data col-
921
+ lected during the baseline data collection period.
922
+ The root causes of these errors were traced to
923
+ cognitive mechanisms such as limited attentional
924
+ resources and working memory limitations. By
925
+ developing an understanding of the pattern and
926
+ types of errors most frequently observed in this
927
+ task domain we hope to provide a DSS which will
928
+ mitigate these errors.7 DiscussionFailure to take appropriate actions may be ex-plained by the limited resource capacity of human
929
+ memory. In these scenarios a large number of
930
+ contacts are monitored for changes in any of sev-
931
+ eral key parameters. Three modules in the DSS
932
+ are hypothesized to assist with recognizing a
933
+ problem and taking the appropriate actions: track
934
+ history; response manager; and the track priority
935
+ list and alerts.Features offered by the DSS to address errorsattributed to limited attentional resources include
936
+ focusing attention on (1) high priority contacts
937
+ (i.e., track priority list and alerts), as well as on (2)
938
+ missing data (e.g., basis for assessment), and (3)
939
+ enabling the decision maker to use more data than
940
+ is typically used in current systems (e.g., track
941
+ history, comparison to norms). Current systems
942
+ require the user to retain previous contact data in
943
+ memory to compare with current values for criti-
944
+ cal parameters. Current systems also require the
945
+ user to rely on recall of vast amounts of informa-
946
+ tion from training and experience. Presenting all
947
+ known data on a contact in a synthesized way
948
+ should reduce working-memory requirements and
949
+ facilitate recognition. Additional potential per-
950
+ formance enhancement features, offered by theDSS, include displaying the complete kinematiccontact history, presenting graphic displays of lo-
951
+ cation and trends, highlighting missing data, pro-
952
+ viding alerts, and providing assessments of cur-
953
+ rent contact identity that go beyond what existing
954
+ systems currently present.Focusing the user's attention on trend andhistory data should decrease the cognitive work-
955
+ load imposed by these scenarios where many
956
+ contacts must be identified and responded to un-
957
+ der severe time constraints. Similarly, delineating
958
+ trend and history data can assist in the identifica-
959
+ tion of a contact where noticing changes in critical
960
+ parameters is essential. Presentation of trend and
961
+ history data, as well as threat assessment and
962
+ comparison to norms, should also mitigate cogni-
963
+ tive "tunnel vision" effects where the decision
964
+ maker attends to a smaller number of cues when
965
+ under stress.The notion of time is an important character-istic of situation awareness (Harwood, Barnett,
966
+ and Wickens, 1988). The past is critical to under-
967
+ standing the present, and both past and present
968
+ information must be used to predict future events
969
+ (Shrestha, et al, 1995). Endsley (1988) referred to
970
+ the "projection of their (perceived elements) status
971
+ in the near future" when discussing situational
972
+ awareness. However, Endsley also noted the task
973
+ of attending to incoming information and subse-
974
+ quently predicting future events places a heavy
975
+ load on working memory. Several decision sup-
976
+ port modules were developed to assist the user in
977
+ remaining aware of the contact's history and
978
+ changes over time. Remembering which actions
979
+ are to be taken at what time levies an additional
980
+ burden by placing a heavy load on working mem-
981
+ ory. A secondary time savings should be achieved
982
+ by the DSS acting as an intelligent "advisor," that
983
+ is, by assisting the decision maker in knowing
984
+ what actions to take, when to take them, and
985
+ which actions have already been taken. A tertiary
986
+ time savings can be achieved by including a tem-
987
+ plate depicting the weapons' release ranges so the
988
+ decision maker does not need to rely on fallible
989
+ human memory or query a team member regard-
990
+ ing weapons ranges. By graphically depicting a
991
+ synthesized view of a contact's kinematic history,
992
+ with the focus on changes in the contact's behav-
993
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 15Monterey, Ca.ior over time, along with the contact's weaponsenvelope in relation to both own-ship's radar and
994
+ weapon coverage, information processing time
995
+ can be saved for the decision maker.8 ConclusionsThe research reported here focuses on developinga DSS which reflects the natural decision-making
996
+ strategies of humans. Presenting synthesized in-
997
+ formation in the form of graphic presentations is
998
+ expected to reduce the cognitive processing load
999
+ for the decision maker when performing situation
1000
+ assessment. The intention is to aid the decision
1001
+ maker by providing information in a way that will
1002
+ minimize the need to maintain information in
1003
+ working memory, reduce information processing
1004
+ demands, help focus attentional resources on the
1005
+ highest priority contacts, remind the user of ac-
1006
+ tions which need to be taken, help make decisions
1007
+ under stress, and support higher levels of situation
1008
+ awareness.Decision support systems require that the hu-man's strengths be used in synergy with the ad-
1009
+ vantages offered by the DSS. Limitations associ-
1010
+ ated with the current generation of automated de-
1011
+ cision aids include the idea that (1) they cannot
1012
+ adequately capture the expertise developed by ex-
1013
+ perience over time and (2) since all contingencies
1014
+ cannot be anticipated, the expert's abilities to use
1015
+ intuition is indispensable (Mosier, in press). Mo-
1016
+ sier's review of the limitations of automated deci-
1017
+ sion systems delineates the characteristics of hu-
1018
+ man expertise that surpass the capabilities of
1019
+ automated systems. These include the human ca-
1020
+ pacity for creativity, adaptability, the ability to
1021
+ incorporate experience, the presence of a broad
1022
+ focus, analogical reasoning, and commonsense
1023
+ knowledge. The goal for the DSS is to capitalize
1024
+ on the strengths of the human along with the ad-
1025
+ vantages provided by the decision support system.9 Testing the DSSThe prototype DSS display modules are currentlybeing empirically evaluated in the simulated tacti-
1026
+ cal environment provided in the DEFTT Labora-
1027
+ tory. Experienced naval decision makers engage
1028
+ in experimental scenarios with and without access
1029
+ to the DSS display. The various decision supportmodules will be tested individually and in combi-nation in future experiments. Data on reduction of
1030
+ errors, improvements in users' situation awareness
1031
+ scores, changes in communication patterns, and
1032
+ subjective responses to the decision support sys-
1033
+ tem will be collected.10 Future ResearchWhile tools based on both the RPD and explana-tion-based reasoning models of decision making
1034
+ are included in the DSS there is no direct connec-
1035
+ tion between the two. Research is currently being
1036
+ conducted to extend schema theory to dynamic
1037
+ decision-making situations. This involves devel-
1038
+ oping and testing a hybrid model of cognitive be-
1039
+ havior in decision making to incorporate both
1040
+ types of knowledge, i.e., feature matching and
1041
+ story generation, as elements of the same schema
1042
+ model of naturalistic decision making (Smith &
1043
+ Marshall, in press). Schema theory as described
1044
+ by these authors, offers a context for integrating
1045
+ these two models which have typically been
1046
+ viewed as separate entities.AcknowledgmentsThe authors gratefully acknowledge the assistanceof Steve Francis, Brent Hardy, C.C. Johnson, Pat
1047
+ Kelly, Ron Moore, Connie O’Leary, Pat Marvel,
1048
+ Mike Quinn, and Will Rogers in data collection,
1049
+ interpretation, developing the DSS, and in con-
1050
+ ducting this research.ReferencesAdams, M. J., Tenney, Y. J., & Pew, R. W.(1995). Situation Awareness and the Cognitive
1051
+ Management of Complex Systems. HumanFactors, 37(1), 85-104.American Psychological Association, ScienceAgenda, Fall, 1988.Bower, G. H. (1970). Organizational Factors inMemory. Cognitive Psychology, 1, 18-46.Broadbent, D. E. (1957). A mechanical model forhuman attention and immediate memory. Psy-chological Review, 64, 205-215.Buck, L. (1989). Human error at sea. HumanFactors Bulletin, September, 32 (9), 12.
1052
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 16Monterey, Ca.Casey, S. (1993). Set Phasers on Stun and OtherTrue Tales of Design, Technology, and Hu-
1053
+ man Error. Aegean Publishing Company,Santa Barbara, CA.Casner, S. (1991). Task-Analytic Design ofGraphic Presentations. Technical Report AIP-145. Departments of Computer Sci-ence and Psychology, Carnegie Mellon Uni-
1054
+ versity and Learning Research and Devel-
1055
+ opment Center, University of Pittsburgh, PA.Cohen, M. S. (1993). The Bottom Line: Natu-ralistic Decision Aiding. In G. Klein, J. Ora-
1056
+ sanu, R. Calderwood, & C. E. Zsambok (Eds.)Decision Making in Action: Models andMethods (pp. 138-147). Ablex PublishingCorporation, New Jersey.Dawes, R. M. (1971). A case study of graduateadmissions: Application of three principles of
1057
+ human decision making. American Psycholo-gist. 34, 571-582.Dotterway, K. A. (1992). Systematic Analysis ofComplex Dynamic Systems: The case of the
1058
+ USS Vincennes. Unpublished master's thesis,Naval Postgraduate School, Monterey, CA.Endsley, M. R. (1988). Design and evaluation forsituational awareness enhancement. Pro-ceedings of the Human Factors 32nd Annual
1059
+ Meeting, 97-101.Federico, P. A. (1995). Expert and novice recog-nition of similar situations. Human Factors,37(1), 105-122.Formal Investigation into the Circumstances Sur-rounding the Downing of a Commercial Air-
1060
+ line by the USS Vincennes, by Rear Admiral
1061
+ William M. Fogerty, USN, 28 July 1988, pp.
1062
+ 4, 6. Hereafter referred to as the Fogerty Re-
1063
+ port.Gruner, W. P. No Time For Decision Making.(1990, November). U.S. Naval Institute Pro-ceedings, 39-41.Harwood, K., Barnett, B., and Wickens, C.(1988). Situational awareness: A conceptual
1064
+ and methodological framework. Proceedingsof the Symposium Psychology in the Depart-
1065
+ ment of Defense.Hockey, G. R. (1986). Changes in Operator Effi-ciency as a Function of Environmental Stress,
1066
+ Fatigue, and Circadian Rhythms. In: K. R.Boff, L. Kaufman, J. P. & Thomas (Eds.):Handbook of Perception and Human Per-formance. Wiley, New York.Hutchins, S. G. (in press). Decision-MakingEvaluation Facility for Tactical Teams. Naval
1067
+ Command, Control, and Ocean Surveillance
1068
+ Center, RDT&E Division Technical Report, in
1069
+ press, San Diego, CA.Hutchins, S. G., Kelly, R. T. and Morrison, J. G.(in press). Decision Support for Tactical Deci-
1070
+ sion Making Under Stress. Proceedings of theSecond International Symposium on Com-
1071
+ mand and Control Research and Technology.June 1996, Monterey, CA.Hutchins, S. G. and Kowalski, J. T. (1993). Tac-tical Decision Making Under Stress: Prelimi-
1072
+ nary Results and Lessons Learned. Proceed-ings of the 10th Annual Conference on Com-
1073
+ mand and Control Decision Aids. June 1993,Washington, D. C.Hutchins, S. G. and Rummel, B. K. (1995). ADecision Support System for Tactical Deci-
1074
+ sion Making Under Stress. Proceedings of the
1075
+ First International Conference on Command
1076
+ and Control Research and Technology. June
1077
+ 1995, Washington, D. C.Hutchins, S. G. and Westra, D. P. (1995). Patternsof Errors Shown by Experienced Navy Com-
1078
+ bat Information Center Teams. Proceedingsof the 39th Annual Meeting of the Human
1079
+ Factors and Ergonomics Society, San Diego,CA. October 1995.Hutchins, S. G. and Westra, D. P. (in preparation).TADMUS Baseline Experimental Results.
1080
+ Naval Command, Control, and Ocean Sur-
1081
+ veillance Center, RDT&E Division Technical
1082
+ Report, in preparation, San Diego, CA.Kaempf, G. L. and Militelo, L. G., (1992). Deci-sion Making in Emergencies, First Offshore
1083
+ Installation Management Conference: Emer-
1084
+ gency Command Responsibilities Collected
1085
+ Papers, Aberdeen, Scotland.Kaempf, G. L. , Wolf, S. and Miller, T. E. (1993).Decision Making in the Aegis Combat Infor-
1086
+ mation Center. In Proceedings of the Human
1087
+ Factors and Ergonomics Society 37th Annual
1088
+ Meeting (pp. 1107-1111). Santa Monica, CA:
1089
+ Human Factors and Ergonomics Society.
1090
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 17Monterey, Ca.Kelly, R. T. , Hutchins, S. G., and Morrison, J. G.(1996). Decision Processes and Team Com-
1091
+ munications with a Decision Support System.Proceedings of the Second International Sym-posium on Command and Control Research
1092
+ and Technology. June 1996, Monterey, CA.Kelly, R. T. , Morrison, J. G., and Hutchins, S.G., (in press). Impact of Naturalistic Decision
1093
+ Support on Tactical Situation Awareness.Proceedings of the 40th Annual Meeting of theHuman Factors and Ergonomics Society,Philadelphia, PA. September 1996.Kirshenbaum, S. S. (1992). Influence of Experi-ence on Information-Gathering Strategies.Journal of Applied Psychology, 77, 343-352.Klein, G. A. (1989). Recognition-Primed Decisions. In W. R. Rouse (Ed.) Advancesin Man-Machine Systems Research (pp. 47 - 92), Vol. 5. JAI Press, Inc.Klein, G. A. (1993). A Recognition-Primed Deci-sion (RPD) Model of Rapid Decision Making.
1094
+ In G. A. Klein, J. Orasanu, R. Calderwood, &
1095
+ C. E. Zsambok (Eds.) Decision Making inAction: Models and Methods (pp. 138-147).Ablex Publishing Corporation, New Jersey.Larkin, J. H. and Simon, H. A. (1987). Why adiagram is (sometimes) worth 10,000 words.Cognitive Science, 1, 65-99.Mosier, (in press). Myths associated with Auto-mated Decision Aids. In G. A. Klein, & C. E.
1096
+ Zsambok (Eds.) Advances in Naturalistic De-cision Making: Research and Applications,Hillsdale, NJ: Erlbaum.Mundy, C. E., Jr. (1994). Thunder and Light-ning: Joint Littoral Warfare, Joint ForceQuarterly, 4, Spring, 45-50.Noble, D. (1989). Application of theory of cog-nition to situation assessment. Vienna, VA:Engineering Research Associates.Noble, D. (1993). A Model to Support Devel-opment of Situation Assessment Aids. n G.
1097
+ A. Klein, J. Orasanu, R. Calderwood, & C. E.
1098
+ Zsambok (Eds.) Decision Making in Action:Models and Methods (pp. 287-305). AblexPublishing Corporation, New Jersey.Norman, D. A. (1988). The Psychology of Eve-ryday Things. Basic Books, Inc. New York.Norman, D. A. and Bobrow, D. G. (1975). Onthe data-limited and resources-limited proc-
1099
+ esses. Cognitive Psychology, 7, 44-64.Office of Naval Technology. (1992). FY 1993Program Plan for Tactical Decision-Making
1100
+ Under Stress, Arlington, VA: July 1992.Paradies, M. (1991). Root Cause Analysis andHuman Factors. Human Factors Society Bul-letin, 34(8), 1-4.Paradies, M. & Unger, L. (1991). TapRoot Inci-dent Investigation System Manual. Volumes1-7. System Improvements, Inc. Knoxville,
1101
+ TN.Pennington, N. & Hastie, R (1992). Explainingthe Evidence: Tests of the Story Model of De-
1102
+ cision Making. Journal of Personality and So-cial Psychology, Vol. No. 2, 189-206.Pennington, N. & Hastie, R (1993). A theory ofExplanation-Based Decision Making. In G. A.
1103
+ Klein, J. Orasanu, R. Calderwood, & C. E.
1104
+ Zsambok (Eds.) Decision Making in Action:Models and Methods (pp. 188-201). AblexPublishing Corporation, New Jersey.Perrow, C. (1984). Normal Accidents: Livingwith High Risk Technologies. New York, Ba-sic Books, Inc.Rasmussen, J. (1986). Information Processingand Human-Machine Interaction. In A. P.
1105
+ Sage (Ed.) Series Volume 12, North-Holland,
1106
+ Amsterdam.Roberts, N. C. & Dotterway, K. A. (1995). TheVincennes Incident: Another Player on theStage? Defense Analysis Vol 11, No. 1, pp.31-45.Salthouse, T. A. (1992). Cognition and Context.Science, 257, 982-983.Sarter, N. B. and Woods, D. D. (1991). Situationawareness: A critical but ill-defined phe-
1107
+ nomenon. The International Journal of Avia-tion Psychology, 1(1), 45-57.Schneider, W., and Detweiler, M. (1988). Therole of practice in dual-task performance:
1108
+ Toward workload modeling in a connectionist/
1109
+ control architecture. Human Factors, 30, 539-566.Senders, J. W. & Moray, N. P. (1991). HumanError: Cause, Prediction, and Reduction.Lawrence Erlbaum Associates, New Jersey.
1110
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 18Monterey, Ca.Shrestha, L. B., Prince, C., Baker, D. P. and Salas,E., (1995). Understanding Situation Aware-
1111
+ ness: Concepts, Methods, and Training. Hu-man/Technology Interaction in Complex Sys-
1112
+ tems. Vol 7, W. B. Rouse (Ed.). San Fran-cisco: JAI Press.Simon, H. A. (1978). Information-processingtheory of human problem solving. In W. K.
1113
+ Estes (Ed.), Handbook of Learning and Cog-nitive Processes, Vol 5, Human InformationProcessing, New York: Wiley.Smith, D. E. and Grossman, J. D. (1993). Under-standing and Aiding Decision Making in
1114
+ Time-Constrained and Ambiguous Situations.Unpublished manuscript.Smith, D. E. and Marshall, S. (in press). Apply-ing Hybrid Models of Cognition in Decision
1115
+ Aids. In G. A. Klein, & C. E. Zsambok (Eds.)Advances in Naturalistic Decision Making:Research and Applications, Hillsdale, NJ:Erlbaum.Swets, J. A. (1984). Mathematical models of at-tention. In R. Parasuraman and R. Davies
1116
+ (Eds.), Varieties of Attention (pp. 183-242).New York: Academic.Tolcott, M. A. (1991). Understanding and AidingMilitary Decisions. Office of Naval ResearchEuropean Office. 27th International Applied
1117
+ Psychology Symposium, Stockholm, Sweden,
1118
+ June 1991.Wilson, G. L. & Zarakas, P. (1978). Anatomy ofa blackout. IEEE Spectrum, February, 339-346.
1119
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 19Monterey, Ca.The requirement to interleave a mul-tiplicity of tasks—although not necessary
1120
+ an ongoing characteristic of shipboard
1121
+ scenarios—represents the type of situation
1122
+ where providing decision support may
1123
+ make the critical difference in the outcome
1124
+ for a scenario. For example, during the
1125
+ experimental scenarios the decision mak-
1126
+ ers may have to perform the following: monitor ship location (relative toother ships and objects in the vicinity) monitor and apply rules of en-gagement to all applicable tracks in the
1127
+ local operating area
1128
+ 
1129
+  receive and send radio messagesto the battle group commander and other operating units in the area
1130
+  monitor tracks on the Aegis dis-play system and maintain situation
1131
+ awareness for all contacts of
1132
+ interest monitor performance of actionstaken by team members to assess the situation
1133
+  monitor the tactical action offi-cer's/commanding officer's performance maintain communications withCIC team members regarding their assessment of tracks and vari-ous actions taken
1134
+ In broad perspective, although teammembers spend much of their time in rou-
1135
+ tine activities, a number of different atten-
1136
+ tionally demanding, knowledge-intensive,
1137
+ and procedurally complex tasks may de-
1138
+ mand attention at any moment. Each of
1139
+ these tasks is usually triggered by a
1140
+ stimulus event, such as a communication
1141
+ from a team member or an alert, and, in
1142
+ order to obtain proper interpretation, mayrequire additional information-seeking be-havior. The cognitive challenge of select-
1143
+ ing and interpreting information to main-
1144
+ tain and revise one's SA is inherently
1145
+ complex. (Jager, Tenney, and Pew, 1995
1146
+ HF) Problems arise when, in the dynamic
1147
+ and multidimensional environments of
1148
+ some littoral antiair warfare scenarios, the
1149
+ situation-critical data become more time-
1150
+ compressed or ambiguous than humans
1151
+ can handle within the inherent time con-
1152
+ straints of the evolving scenario.Resources are such things as processing effort,the various forms of memory capacity, and
1153
+ communications channels (Bobrow & Norman,
1154
+ 1975, in Rasmussen et al).Topics to be discussed include: (1) a descriptionof the difficult tasks identified for analysis; (2) the
1155
+ general methodological approach; (3) develop-
1156
+ ment of the performance measures and issues
1157
+ related to their development; (4) discussion of the
1158
+ ; and (5) discussion of the types of errors made by
1159
+ decision makers and interpretations for the cause
1160
+ of these errors based in the cognitive psychology
1161
+ literature.
1162
+ Paper presented to the Second International Command and Control Research and Technology Symposium,20Monterey, Ca.tactical operations require decision makingconditions of time pressure, stress, am-
1163
+ biguous, inaccurate and missing informa-
1164
+ tion, uncertain communications, and
1165
+ shifting conditions. These conditions make
1166
+ it difficult to perform careful analysis prior
1167
+ to making decisions. Traditionally, most
1168
+ decision research and decision support
1169
+ system development have focused on well-
1170
+ defined tasks in carefully controlled envi-
1171
+ ronments. Recent research has suggested
1172
+ that tactical decision makers use experi-
1173
+ ence to generate a likely course of action
1174
+ and then evaluate its feasibility using
1175
+ mental simulation.One way to prevent the same type of casualtyfrom being repeated is to thoroughly investigate
1176
+ and analyze root causes of actual mishaps, as well
1177
+ as data collected in a simulated tactical
1178
+ environment, and apply the findings in a concrete
1179
+ manner to improve tactical decision making.According to the recognition-primed decision-making (RPD) model, experienced decision
1180
+ makers can make rapid, high-quality decisions by
1181
+ associating a situation directly with the actions
1182
+ that normally work well in that kind of situation.
1183
+ Roberts, K. H., Stout, S. K., and Halpern, J. J.(1994) Decision Dynamics in Two High
1184
+ Reliability Military Organizations. Man-agement Science, Vol. 40, No. 5, May1994, 614-624.Tetlock, P. E., Accountability and the Persever-ance of First Impressions, Social Psycho-logical Quarterly, 26(1983), 285-292.Tetlock, P. E., Accountability: The NeglectedSocial Context of Judgment and Choice, in
1185
+ L. L. Cummings and B. M. Straw (Eds.),
1186
+ Research in Organizational Behavior, Vol.7, JAI Press, Greenwich, CT, 1985, 297-332.The comparison to norms tool is based on acognitive model of human information processing
1187
+ which uses a feature matching strategy. The
1188
+ model proposes a data-driven process. As such,
1189
+ the comparison to norms tool is a knowledge-
1190
+ based tool with the knowledge represented as
1191
+ templates. Each template is a linked timeline
1192
+ associating individual features, through a series of
1193
+ feature matches, with expected actions. This
1194
+ allows the tool to make an assessment of the
1195
+ situation presented to the decision maker. A
1196
+ related module is the response manager. This
1197
+ assessment consists of a categorization of the
1198
+ situation (e.g., “hostile aircraft attacking”) and
1199
+ presentation of an associated template. The
1200
+ response manager module shows a template for an
1201
+ assessment of the current situation. This is a
1202
+ timeline display for template features, or events.
1203
+ The response manager module also provide inputs
1204
+ to the Prioritized track list, Alerts, and Responses
1205
+ and Tripwires.SABER is a model of another cognitivestrategy employed in making decisions. This
1206
+ strategy is known as explanation-based reasoning,
1207
+ or story generation. In this approach, available
1208
+ data are assembled into explanatory structures,
1209
+ with one structure for each possible conclusion.
1210
+ Each of the explanations attempts to explain how
1211
+ every piece of data can be accounted for in
1212
+ support of each conclusion, even though some of
1213
+ the data items would naturally contradict reaching
1214
+ some conclusions. Contradictory data are
1215
+ explained through the of internal assumptions. It
1216
+ is assumed that there are a fixed number of pre-
1217
+ defined possible conclusions and each data item
1218
+ points directly to one of those possible conclu-
1219
+ sions.Once the explanations are constructed,SABER evaluates them to determine which seems
1220
+ most plausible. Plausibility is based on three
1221
+ Paper presented to the Second International Command and Control Research and Technology Symposium,21Monterey, Ca.criteria: simplicity, completeness, and impor-tance. SABER provide data products which relate
1222
+ data to explanatory hypotheses for the current
1223
+ situation. Hypotheses are presented in rank order,
1224
+ with evidence for and against each hypothesis and
1225
+ missing data is presented below the corresponding
1226
+ hypothesis. For a more detailed description of the
1227
+ DSS the reader is referred to Hair & Pickslay,
1228
+ 1992; Hair, Pickslay, & Chow, 1992; Hutchins
1229
+ and Rummel, 1994.Hair, D. C. and Pickslay, K. (1992). Explanation-Based Reasoning in Decision Support
1230
+ Systems, Proceedings of the 9th AnnualConference on Command and Control De-cision Aids, June 1992, Monterey, CA.Hair, D. C., Pickslay, K. & Chow, S. (1992)Explanation-Based Decision Support in
1231
+ Real Time Situations. Proceedings of the1992 IEEE International Conference onTools with AI. Nov. 1992, Arlington, VA.Various studies have indicated that as much as 90
1232
+ percent of industrial and system failures are
1233
+ produced by human error (Senders & Moray,
1234
+ 1991).
1235
+ At the same time, a shift was occurring in U.S.
1236
+ Navy doctrine, away from a "blue-water" strategy
1237
+ to a doctrine of littoral operations aimed at
1238
+ potentially hostile regional powers.Tactical decision makers in today's oper-ating environment are required to perform
1239
+ complex tasks in a highly dynamic envi-
1240
+ ronment. Numerous interactive surface
1241
+ units and aircraft whose parameters are in
1242
+ flux and which must be continually
1243
+ sensed, processed, their future status pro-
1244
+ jected (development of hypotheses re-
1245
+ garding their future behavior) and actions
1246
+ taken to assure the successful outcome.Operating close to land presents additionalchallenges to the tactical decision maker.
1247
+ In littoral (i.e., near-land) settings, which
1248
+ most likely represent the majority of future
1249
+ anticipated naval conflicts, the decisions to
1250
+ be made are even more complex than they
1251
+ would be in full-scale warfare. When ci-
1252
+ vilian and neutral nation resources are in
1253
+ the conflict area incoming information
1254
+ carries an added element of uncertainty. In
1255
+ these situations, interpretation of the rules
1256
+ of engagement, contact identification, de-
1257
+ termining the capability and possible in-
1258
+ tent of the potential threat, and the
1259
+ shoot/no-shoot decision often pose ex-
1260
+ tremely difficult decision problems. (Since
1261
+ 70 percent of the world's population lives
1262
+ within 200 miles of the sea, most future
1263
+ contingencies are likely to involve littoral
1264
+ warfare (Mundy, 1994).)as opposed to replacing "the user's approach tothe problem" (emphasis in original, Cohen, 1993)with tools based on decision analysis or
1265
+ mathematical optimization,
1266
+ A considerably smaller number are attributable to
1267
+ other causes such as mechanical, electrical and
1268
+ materials failure (Meshkati, 1993). The purpose for this phase of the TADMUSprogram is to empirically evaluate the effective-
1269
+ ness of a DSS based on these recent approaches to
1270
+ decision support, to the extent possible,"This focus on why errors occuris...different from...the typical study of
1271
+ human errors which solely emphasize what
1272
+ occurs, a point of view which has received
1273
+ considerable criticism." (Rouse & Rouse,
1274
+ 1983, p. 539)
36733527-1-23 ADDED
@@ -0,0 +1,1319 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 2Monterey, Ca.tact identification, intent, available responses andtheir consequences. For example, the close prox-
2
+ imity of U.S. Navy forces and potential adversary
3
+ forces makes interpreting the actions of an in-
4
+ bound aircraft who does not respond to radio
5
+ warnings much more difficult. Should the air-
6
+ craft's behavior be interpreted as an attack profile,
7
+ or does the pilot merely intend to harass, or does
8
+ the aircraft in question not carry the equipment
9
+ necessary to receive verbal warnings, leaving the
10
+ pilot unable to receive radio warnings directed
11
+ toward him and unaware of his precarious posi-
12
+ tion? In extreme cases there is no clear cut right or
13
+ wrong answer about a decision. Rapidly unfolding
14
+ events result in severe time pressure and severe
15
+ (often catastro-phic) consequences for errors.
16
+ While current real-time battle management sys-
17
+ tems are well-suited to the demands of all-out
18
+ conflicts, they may not be optimized for littoral
19
+ situations where human intervention in decision-
20
+ making is even more important (Office of Naval
21
+ Technology [ONT], 1992). (Since 70 percent of
22
+ the world's population lives within 200 miles of
23
+ the sea, most future contingencies are likely to
24
+ involve littoral warfare (Mundy, 1994).)Two unfortunate and highly publicized eventsfocused attention on the difficult types of deci-
25
+ sions confronting naval commanders and provided
26
+ the impetus for this research. In the case of the
27
+ U.S.S. Stark, the commander made the decision to
28
+ not engage an inbound aircraft which was be-
29
+ lieved to not be a threat to his ship, and 27 U.S.
30
+ naval personnel lost their lives as a result. In the
31
+ case of the U.S.S. Vincennes, the commander
32
+ made the decision to engage the inbound aircraft
33
+ believing it was a threat to his ship—which turned
34
+ out to be a commercial airliner—and all personnel
35
+ aboard the airliner were killed as a result. In rec-
36
+ ognition of the complex and difficult decisions
37
+ required in these types of situations the Tactical
38
+ Decision Making Under Stress (TADMUS) pro-
39
+ gram was initiated to conduct research in the areas
40
+ of human factors and training technology. The
41
+ objective is to develop and apply principles that
42
+ can help avoid these types of situations in the fu-
43
+ ture. This paper, and a two companion papers
44
+ (Hutchins, Kelly, & Morrison, 1996; Kelly,
45
+ Hutchins, & Morrison, 1996), report on a multi-year, multi-experiment research effort conductedunder the TADMUS program to apply recent de-
46
+ velopments in decision theory and human-system
47
+ interaction technology to the design of a decision
48
+ support system (DSS) for enhancing tactical deci-
49
+ sion making under highly complex conditions.1.1 "Naturalistic" and Classical Decision-Making Para-digmsIn the same time frame that these tragic acci-dents occurred, a radical shift was occurring in the
50
+ way psychologists viewed human decision mak-
51
+ ing. Research was now focused on experienced
52
+ decision makers performing their normal tasks in
53
+ natural settings. "Naturalistic" decision-making
54
+ research studies the decision strategies people ac-
55
+ tually use in bringing their expertise to bear under
56
+ challenging real-world conditions. Decision
57
+ making milieus encompassed under the naturalis-
58
+ tic paradigm include hospital emergency rooms,
59
+ aircrew flight coordination, military command and
60
+ control settings, process control systems, and po-
61
+ lice and fire units.The naturalistic perspective, also known as"everyday cognition," is based on a belief that
62
+ cognitive functions "elicited in natural settings
63
+ (are) likely to differ, either quantitatively, or
64
+ qualitatively, from those that occur in artificial or
65
+ contrived situations, and results from sterile and
66
+ contrived situations may not generalize to less
67
+ constrained and more natural environments"
68
+ (Salthouse, 1992, p. 982). There is a growing
69
+ body of work that demonstrates that experienced,
70
+ real-world decision makers rarely use traditional
71
+ resource intensive strategies to make decisions in
72
+ the face of dynamic, adverse conditions and time-
73
+ pressure (Kaempf & Militelo, 1992; Klein, 1989;
74
+ 1993). Instead, experts rely on their abilities to
75
+ recognize and appropriately classify situations:
76
+ these abilities are based on having much experi-
77
+ ence in the task domain. Once they know what
78
+ they are facing they also tend to know what re-
79
+ sponse option to apply, based on retrieval from
80
+ memory of typical responses and outcomes that
81
+ worked well in past similar situations. They use
82
+ the limited time available to evaluate the feasibil-
83
+ ity of that option before implementing it. Experi-
84
+ enced decision makers recognize the situation or
85
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 3Monterey, Ca.scenario based on a comparison of the features ofthe current situation with stored memory represen-
86
+ tations, or schemata. Schemata are highly inter-
87
+ con-nected clusters of knowledge concerning cer-
88
+ tain situations, or particular problem types, and
89
+ associated actions or solution procedures (Fede-
90
+ rico, 1995). Once the situation is recognized, so-
91
+ lutions are stimulated by activation of these mem-
92
+ ory representations.In contrast to the naturalistic perspective, ear-lier analytical methods applied in decision support
93
+ systems were primarily used for option generation
94
+ and evaluation, rather than for situation assess-
95
+ ment. Traditional decision theorists argue that op-
96
+ timum decision making involves thorough analy-
97
+ sis of all the available data and the evaluation of
98
+ all possible hypotheses; these approaches tend to
99
+ rely on extensive calculations designed to arrive at
100
+ optimal solutions. Processes for making decisions
101
+ where the person weighs the pros and cons of
102
+ various options and selects the one option that
103
+ provides the most benefit are described in the de-
104
+ cision making literature. The extensive time re-
105
+ quirements and complicated mathematical calcu-
106
+ lations involved (e.g., Multi-Attribute Utility
107
+ Analysis), however, make these approaches unre-
108
+ alistic for situations requiring rapid decision
109
+ making. Generally, these analytical strategies in-
110
+ volve the following steps (Kaempf & Militello,
111
+ 1992): specify all relevant features of the task;
112
+  identify the full range of options;
113
+  identify the key evaluation dimensions;
114
+  identify weights for each dimension;
115
+  rate each option on each dimension;
116
+  tabulate the results, and
117
+  select the best option.These analytical strategies may be appropriate for
118
+ inexperienced subjects making decisions about
119
+ novel tasks, but not for experienced personnel
120
+ making real-time decisions. In natural settings,
121
+ time constraints and the difficulty in assigning
122
+ weights and rating the dimensions involved render
123
+ classical analysis techniques untenable. Typically
124
+ in realistic settings experts employ recognition-
125
+ based reasoning, not classical analytical ap-
126
+ proaches. Experienced decision makers use their
127
+ extensive knowledge to seek information, identifyand interpret the problem, understand the signifi-cance, derive the intention (where possible),
128
+ model the situation (as time allows), select the
129
+ action, evaluate the choice, and anticipate the con-
130
+ sequence. "This decision cycle is distinctly differ-
131
+ ent from classical models, which are based on the
132
+ assumption that all options, outcomes, and prefer-
133
+ ences are known and calculated in advance"
134
+ (Federico, 1995, p. 106).Traditionally, research in decision making hasbeen directed largely toward situations in which
135
+ (1) decision makers have sufficient time to gener-
136
+ ate options, conduct option assessment, and select
137
+ a course of action; (2) the consequences of an in-
138
+ correct response are not immediately severe; (3)
139
+ decisions are reached via consensus of a group;
140
+ and (4) workload is manageable. Little research
141
+ has been conducted into the development of tacti-
142
+ cal decision support systems for use in naturalistic
143
+ situations characterized by time pressure, high
144
+ risk, uncertainty and information ambiguity, high
145
+ workload, team coordination demands and task
146
+ complexity (ONT, 1992).The central hypothesis for the research re-ported here is that presenting decision makers
147
+ with decision support tools which were designed
148
+ to parallel the cognitive strategies employed by
149
+ experts, as observed in naturalistic settings, will
150
+ reduce the number of decision making errors. This
151
+ is accomplished by developing the architecture
152
+ and algorithms to process information the same
153
+ way research indicates humans do under similar
154
+ circumstances.2 Tactical Decision Making TasksThe global tactical decision-making task involvesidentification of and responding to numerous
155
+ contacts. When an aircraft (or a surface contact) is
156
+ detected the CIC personnel work as a team to de-
157
+ termine the identity and to try to determine
158
+ whether or not the aircraft poses a threat. The high
159
+ degree of inherent ambiguity associated with
160
+ contact information can often make threat assess-
161
+ ment a very difficult task. This is because many
162
+ pieces of data fit multiple hypotheses regarding
163
+ threat assessment. The global response choices
164
+ (that is, engage, monitor, do nothing) are largely
165
+ determined by the ship's orders and the current
166
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 4Monterey, Ca.geopolitical situation. Specific actions (such as,change course, issue verbal warnings, illuminate
167
+ with radar, challenge with other sensors, etc.) de-
168
+ pend on the local conditions and the relative posi-
169
+ tions of the inbound contact of interest and own-
170
+ ship. Determining which of these actions is likely
171
+ to be effective depends on maintaining an accurate
172
+ threat assessment which requires "continual up-
173
+ dating in accordance with recurrent situation as-
174
+ sessments" (Sarter & Woods, 1991, p. 52).This decision problem presents a highly chal-lenging cognitive task, that is, making inferences
175
+ and deductions from incomplete and uncertain
176
+ information derived from multiple sources and
177
+ relating to several concurrent threats (or potential
178
+ threats) under time-compressed conditions. The
179
+ cognitive functions performed by the tactical deci-
180
+ sion maker are both data and resource limited
181
+ (Norman & Bobrow, 1975). Decisions are re-
182
+ source limited by the mental resources of the deci-
183
+ sion makers, who must maintain large amounts of
184
+ information in memory under conditions of high
185
+ workload and stress. The decisions are data lim-
186
+ ited by the inability of the sensors to provide
187
+ complete, error-free, unambiguous data to support
188
+ the identification process. In particular, the ex-
189
+ perimental scenarios were designed to follow the
190
+ pattern of being set in an ambiguous situation
191
+ where one or more threats of uncertain origin and
192
+ uncertain intent approach either own ship or the
193
+ ship being protected and may not respond to
194
+ warnings. Scenarios were designed to be highly
195
+ ambiguous, as this quality of uncertainty is in-
196
+ dicative of the types of decisions to be made in
197
+ current and future scenarios.2.1 Threat AssessmentIn the antiair warfare problem, threat assess-ment is particularly difficult because the available
198
+ information is often incomplete or ambiguous.
199
+ The ambiguity could be due to (a) the information
200
+ transmission characteristics of the transmission
201
+ medium, such as a radar transmission or a radio
202
+ report that is only intercepted on an intermittent
203
+ basis, (b) deliberate deceptive actions (such as ra-
204
+ dar jamming) by the pilot flying the aircraft, or (c)
205
+ the overlapping classification categories typical of
206
+ many parameter measure-ments. For example, air-craft can typically fly at altitudes ranging between2,000 and 40,000 feet. Generally, an aircraft that
207
+ is flying above 20,000 feet is not considered to be
208
+ a threat. Conversely, an aircraft below 10,000 ft is
209
+ considered to be more of a potential threat. How-
210
+ ever, aircraft flying in the middle range, (that is,
211
+ below 20,000 ft. and above 10,000 ft.) can be
212
+ much more difficult to categorize. Because many
213
+ aircraft do fly in this middle range, other variables
214
+ need to be considered in conjunction with altitude.
215
+ This same situation of overlapping categorization
216
+ categories exists for several other variables. These
217
+ variables include radars that are found on both
218
+ threat and non-threat platforms, country of origin,
219
+ and measures of course and speed. In the case of
220
+ speed, for example, when an aircraft flying at a
221
+ low altitude decreases speed this could be viewed
222
+ as indicative of a threat action (that is, slowing
223
+ down in order to obtain better targeting informa-
224
+ tion); however, at the same time, there could be
225
+ other viable explanations for an aircraft's de-
226
+ creasing speed.If the decision maker had access to all dataabout a contact approximately twelve variables
227
+ would be used to determine identity and to infer
228
+ intent. Two or three of these items, alone, do not
229
+ provide definitive answers because, in many
230
+ cases, these parameter values do not fall within
231
+ clear-cut ranges for a particular assessment cate-
232
+ gory (i.e., threat, non-threat). Thus, a single time
233
+ slice of information provides an incomplete pic-
234
+ ture of the situation. In the dynamic, ambiguous
235
+ conditions characteristic of littoral operations, the
236
+ rate and direction of change (data history) can
237
+ help one better assess the threat and predict the
238
+ future state of the situation. When the incoming
239
+ information changes over time, the integration of
240
+ information as it changes can help the user extract
241
+ the message (Kirshenbaum, 1992). The DSS was
242
+ designed to do precisely this: to facilitate the inte-
243
+ gration process and present a synthesized picture
244
+ of the situation to the user in a format that can be
245
+ quickly assimilated. The variables, that are used to
246
+ develop a threat assessment, can be divided into
247
+ two classes: sensor information (raw or computer-
248
+ processed information) and the contact's response,
249
+ or lack of response, to actions taken by the team.
250
+ These other actions, and the integration of the in-
251
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 5Monterey, Ca.formation received via the contact's response orlack of response to them, are necessary to clarify
252
+ the tactical picture.2.2 Situation AwarenessWhile recent increasing interest among re-searchers regarding the concept of situation
253
+ awareness (SA) has generated a debate on the pre-
254
+ cise definition of this term most researchers ac-
255
+ knowledge the importance of the concept. In gen-
256
+ eral, SA refers to the decision maker's moment-
257
+ by-moment ability to monitor and understand the
258
+ state of the complex system and its environment
259
+ (Adams, Tenney, & Pew, 1995). These authors
260
+ state the essential idea which is that when emer-
261
+ gencies arise, the completeness and accuracy of
262
+ the decision maker's SA are critical to the ability
263
+ to make decisions, revise plans, and manage the
264
+ system. Specific decision-making tasks included
265
+ under SA include the ability to: (1) maintain an
266
+ accurate perception of the surrounding environ-
267
+ ment (both internal and external to the ship); (2)
268
+ identify problems and/or potential problems; (3)
269
+ recognize a need for action; (4) note deviations in
270
+ the mission; and (5) maintain awareness of tasks
271
+ performed (Shrestha, Prince, Baker and Salas,
272
+ 1995). To maintain an accurate SA the decision
273
+ maker should take into account both information
274
+ that is available and that which can be activated
275
+ from memory (Sarter and Woods, 1991).However, a difficulty arises as a result of theheavy workload imposed by this process. When
276
+ the decision maker is faced with several concur-
277
+ rent contacts of interest, all of which have numer-
278
+ ous associated data items (i.e., as many as a
279
+ dozen), some or all of which may change over the
280
+ course of the scenario (e.g., intelligence, active
281
+ radar emitters, various kinematic parameters, etc.)
282
+ memory load easily exceeds human capacity.
283
+ Moreover, changing parameters may impart dif-
284
+ ferent interpretations to what is occurring. In some
285
+ cases the moment-to-moment attentional demands
286
+ of a tactical situation are relentless and unforgiv-
287
+ ing (such as, a terrorist aircraft directly inbound
288
+ toward "own-ship" which can result in "task fixa-
289
+ tion"), sometimes relevant background knowledge
290
+ is unavoidably incomplete (such as, an unfamiliar
291
+ aircraft), and sometimes the decision maker is al-ready thinking and working as hard as possible,even when there are no unanticipated events when
292
+ there is a high contact density (Adams, et al,
293
+ 1995). These instances provide a few illustrations
294
+ of situations that can degrade situation awareness.Complex information gathering and process-ing systems have been designed to aid the deci-
295
+ sion-maker in the past. However, these systems
296
+ often increase the decision-maker’s burden due to
297
+ the inherent system complexity and the failure to
298
+ design them in a way that they will fit the user's
299
+ cognitive processing limitations. Often, these
300
+ systems require operators to perform difficult
301
+ cognitive tasks under heavy workloads. They must
302
+ perceive, synthesize and determine the relevance
303
+ of a continual stream of incoming information,
304
+ often pertaining to several concurrent contacts,
305
+ while projecting future anticipated events and
306
+ making decisions regarding actions to be taken.
307
+ Decision makers must assess, compare, and re-
308
+ solve conflicting information, while making diffi-
309
+ cult judgments, and remembering the status of
310
+ critical contacts along with the contact's response
311
+ to actions taken by the CIC team. These decision-
312
+ making tasks are interleaved with other required
313
+ tasks, such as keeping other team members in-
314
+ formed (both on and off the ship). Furthermore,
315
+ these complex tasks are performed under condi-
316
+ tions where adverse environmental (noise, vibra-
317
+ tion, temperature extremes, etc.) and internal
318
+ stressors (boredom, fatigue, anxiety, and fear) are
319
+ part of the environment.3 Decision Support PrinciplesA case has been made that previous generations ofdecision support systems, which focus primarily
320
+ on solution optimization and base decision sup-
321
+ port on normative models of human decision
322
+ making, are less applicable than a DSS that par-
323
+ allels the cognitive strategies used by domain ex-
324
+ perts in situations characterized by time-
325
+ constrained situations with uncertain and ambigu-
326
+ ous data (Smith & Grossman, 1993). These
327
+ authors point out that rarely, if ever, were earlier
328
+ tactical decision aids intended as psychological
329
+ models of human cognitive behavior. Instead,
330
+ these aids performed "complex and burdensome
331
+ calculations, reducing the work-load on personnel,
332
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 6Monterey, Ca.speeding up the dissemination of information, andproviding more time for command decision mak-
333
+ ing" (Tolcott, 1991, p. 44).3.1 Feature Matching and Story GenerationWe have applied two of these new models ofhuman decision making—which parallel the cog-
334
+ nitive strategies used by domain experts—to the
335
+ design of a DSS for enhancing antiair warfare
336
+ tactical decision making. These two models that
337
+ people use in assessing a situation are feature
338
+ matching and story generation. The feature
339
+ matching model, described by Noble (1989), in-
340
+ volves an organization of memory, or "schemas,"
341
+ and information-processing where decision mak-
342
+ ers use their previous experiences to assess a
343
+ situation and identify promising actions. Incoming
344
+ information is categorized, selected, edited, and
345
+ organized on the basis of a person's general
346
+ knowledge about a domain. Both story generation
347
+ and feature matching occur under conditions
348
+ where a large base of implication-rich, condition-
349
+ ally dependent pieces of evidence must be evalu-
350
+ ated before choosing an alternative from a set of
351
+ prospective courses of action. The feature match-
352
+ ing model applies a spatio-temporal dependence,
353
+ whereas story generation is an example of causal
354
+ dependence. According to the explanation-based
355
+ model, decision makers construct a causal model
356
+ to explain the available evidence (Pennington &
357
+ Hastie, 1993). At the same time, the decision
358
+ maker creates a set of alternatives from which an
359
+ action will be chosen. A decision is made when a
360
+ story is successfully matched to an alternative in
361
+ the choice set. Story generation occurs in complex
362
+ situations where the decision maker may not have
363
+ all the necessary information or when a series of
364
+ facts may appear to contradict each other. The de-
365
+ cision maker must then develop causal links be-
366
+ tween these facts to produce a coherent picture of
367
+ the situation (Klein, 1989; 1993).The explanation-based reasoning model isbased on research which found that jurors develop
368
+ a narrative story to organize trial information
369
+ where causal and intentional relations between
370
+ events are central (Pennington & Hastie, 1992).
371
+ Pennington & Hastie propose four certainty prin-
372
+ ciples—coverage, coherence, uniqueness, andgoodness-of-fit—that govern (i) which story willbe accepted, (ii) which decision will be selected,and the (iii) confidence or degree of certainty withwhich a particular decision will be made. This or-
373
+ ganization of the evidence by the decision maker
374
+ is believed to facilitate evidence comprehension.
375
+ A central component of this model is that the
376
+ story the juror constructs determines the juror's
377
+ decision. This explanation-based decision process
378
+ is employed when the body of evidence relevant
379
+ to a decision is large, complex, and the implica-
380
+ tions of its components are interdependent.Feature matching, also referred to as the rec-ognition-primed decision (RPD) model, "occurs
381
+ when the decision maker recognizes the features
382
+ of the present situation as similar or identical to
383
+ those of a previous situation" (Kaempf & Militelo,
384
+ 1992, p. 6). An adequate match triggers recall of
385
+ information learned about this type of situation:
386
+ (a) plausible goals, (b) critical cues to be moni-
387
+ tored, (c) expectations of what should happen, and
388
+ (d) a course of action that worked in similar situa-
389
+ tions. According to this recent approach, expert
390
+ decision makers may rely on well-developed
391
+ memory representations to guide decision making
392
+ in new (but similar) situations. The RPD model of
393
+ decision making fuses two processes—situation
394
+ assessment and mental simulation (Klein, 1993).
395
+ In the simplest case the situation is recognized as
396
+ familiar or prototypical, using feature matching,
397
+ and the obvious response is implemented. In a
398
+ more complex case the decision maker performs a
399
+ conscious evaluation of the response, using men-
400
+ tal simulation to uncover problems prior to im-
401
+ plementing the response. In the most complex
402
+ case the evaluation reveals flaws requiring modi-
403
+ fication, or the option is judged inadequate and
404
+ rejected in favor of the next most typical reaction.3.2 Situation AssessmentIn general, the overall task of responding toantiair warfare scenarios consists of situation as-
405
+ sessment ("what's going on") and course of action
406
+ selection ("what to do about it"). Recent theories
407
+ of decision making emphasize the importance of
408
+ situation assessment for good decision making in
409
+ naturalistic, event-driven situations. Moreover,
410
+ they stress that decisions regarding actions to be
411
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 7Monterey, Ca.taken are a by-product of developing the situationawareness that precedes action selection. Klein
412
+ (1989) has found that usually the situation itself
413
+ either determines or constrains the response op-
414
+ tions and that experienced decision makers make
415
+ up to 90% of all decisions without considering
416
+ alternatives. If the situation appears similar to one
417
+ that the decision maker has previously experi-
418
+ enced, the pattern will be recognized and the
419
+ course of action is usually immediately obvious.
420
+ On the other hand, if the situation does not seem
421
+ familiar complex RPD will be involved where the
422
+ decision maker adjusts the option after evaluating
423
+ it.Additional evidence was found in the specifictask domain of interest to the TADMUS program
424
+ which added support to these findings on the way
425
+ real-world decision makers make decisions in the
426
+ context of their normal jobs. Research was con-
427
+ ducted to determine decision requirements for
428
+ command-level decision makers in the combat
429
+ information center (CIC) of an Aegis cruiser.
430
+ Analysis of 14 incidents from actual problems re-
431
+ vealed 183 decisions. Of these, 103 concerned
432
+ situation assessments. Results obtained after ana-
433
+ lysts coded these situation assessments indicated
434
+ that decision makers arrived at approximately
435
+ 87% of their situation assessments through feature
436
+ matching and the remaining 13% through story
437
+ generation (Kaempf, Wolf, & Miller, 1993). The
438
+ other eighty decisions that were identified, from
439
+ analysis of the real-world incidents mentioned
440
+ above, involved course of action selection. These
441
+ course of action decisions served a variety of
442
+ functions, although, relatively few were intended
443
+ to end the incident. Twenty were intended as a
444
+ final course of action decision; 14 were imple-
445
+ mented to obtain more information, 22 to manage
446
+ resources, and 24 to put themselves in a more fa-
447
+ vorable tactical position. A recognition-based
448
+ strategy was also used by decision makers to de-
449
+ velop a course of action, accounting for 95% of
450
+ the actions taken in the 14 incidents. The decision
451
+ makers generated and compared multiple options
452
+ in only 5% of the cases. In line with these find-
453
+ ings, the TADMUS program has adopted the po-
454
+ sition that decision aiding systems should assist in
455
+ the decision making process, and focus on aidingthe situation assessment portion of the decision-making task.A DSS was developed to support decision-making processes which research has shown are
456
+ used by decision makers in real-world settings
457
+ (Hutchins, Kelly, & Morrison, 1996). Specifi-
458
+ cally, the DSS parallels the strategies used by ex-
459
+ perienced decision makers to perform situation
460
+ assessment (Nobel, 1989; 1993). This approach to
461
+ supporting the user's intuitive approach to dealing
462
+ with dynamic decision-making situations should
463
+ produce tools that are both more easily understood
464
+ and used, and that more effectively "exploit the
465
+ decision maker's knowledge and expertise that
466
+ might facilitate adaptation to complex, novel
467
+ situations" (Cohen, 1993, p. 265).4 Human-System Interaction PrinciplesThe vast majority of research on human-computerinteraction design has been devoted to character-
468
+ istics of displays that impact human perception,
469
+ such as symbol legibility or detectability, and on
470
+ relatively simple cognitive functions such as
471
+ memory tasks. Fewer efforts have been devoted to
472
+ understanding the effects of the format and man-
473
+ ner in which information is presented on more
474
+ complex levels of human cognition such as deci-
475
+ sion making. Consequently, principles that can be
476
+ applied to the design of the interface between the
477
+ user and a decision support system for the purpose
478
+ of enhancing cognitive changing situations, are
479
+ not available to any significant degree (ONT,
480
+ 1992).4.1 Graphic PresentationsSeveral advantages are offered by graphicpresentations over a text-based presentation for-
481
+ mat (Larkin and Simon, 1987). Graphic presenta-
482
+ tions should (1) reduce the amount of mental
483
+ computation required to perform tasks; and (2)
484
+ allow users to spend less time searching for
485
+ needed information. Casner (1991) elaborated on
486
+ these ideas and found that graphics allow users to
487
+ substitute less demanding perceptual opera-tions
488
+ for more complex logical operations. For exam-
489
+ ple, determining a change in altitude (and the de-
490
+ gree of change) is immediately apparent when the
491
+ user glances at the track history module. (Note
492
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 8Monterey, Ca.that the words contact and track can be used inter-changeably. The reader is referred to Figure 1.)The objective for the track history module isto facilitate the contact identification process by
493
+ providing information that is integrated in a way
494
+ that supports a recognitional decision strategy.
495
+ This module depicts a contact’s speed, altitude,
496
+ course and range on a two-dimensional graphical
497
+ display along with a geometric representation of
498
+ both the contact’s weapon release envelope and
499
+ own-ship’s weapons coverage. A large amount of
500
+ parametric data is portrayed graphically for rapidassimilation by the user. The user can see, at aglance, a synthesized picture of the contact’s be-
501
+ havior. Compare this rather simple perceptual op-
502
+ eration with the more complex logical operation
503
+ involved in current operational systems which re-
504
+ quire the user to recall and subtract numerical val-
505
+ ues for past and current altitudes.Graphics also allow users to omit steps thatare otherwise necessary when a task is performed
506
+ without a graphic. An example of this advantage
507
+ isFigure 1. Decision Support System Display Modules.also illustrated in the track history module whichincludes templates indicating weapon's coverage
508
+ for both the inbound contact and "own-ship." To
509
+ determine whether the aircraft is within its
510
+ weapon's launch range there is no need to recall
511
+ the specific launch range values and then compare
512
+ them with the aircraft's current range. Instead, the
513
+ user can determine if the aircraft is within its
514
+ launch range by a quick glance at the display.Graphics help users save time when searchingfor needed information when several related di-
515
+ mensions of information are encoded in a single
516
+ graphical object. This is accomplished by inte-
517
+ grating the kinematic parameters of speed, course,
518
+ altitude, bearing, and range for a contact. The user
519
+ can see, at a glance, a synthesized picture of the
520
+ contact's behavior. Compare this process with
521
+ reading, in a text-based format, the individual pa-
522
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 9Monterey, Ca.rameters which need to be integrated by the userinto a coherent picture of the contact's behavior.5 Limited Cognitive Processing CapabilitiesSince there are limits to the cognitive processingcapability of humans, it is important for the sys-
523
+ tem to provide the needed information in a format
524
+ that best supports the user operating under dy-
525
+ namic decision-making conditions. It may be the
526
+ case that current systems are inadequate to support
527
+ the cognitive processing demands required by
528
+ certain littoral scenarios. For example, according
529
+ to Gruner (1990, p. 41), the U.S.S. Vincennes of-
530
+ ficers and system operators "could not make better
531
+ decisions because they did not have time to con-
532
+ firm or deny the information uncertainties pre-
533
+ sented them." Gruner maintains that the rapid pace
534
+ involved in these types of situations can exceed
535
+ the capacity of the human to comprehend the
536
+ rapid flow of information presented by complex
537
+ systems. In the case of the Vincennes, the CIC
538
+ team had three minutes and 40 seconds to make
539
+ their decision. This includes the time required for
540
+ the operators to perceive and interpret sensor data
541
+ and for the commanding officer to make informed
542
+ judgments from these data (Roberts & Dotterway,
543
+ 1995). The result of the human's limited cognitive
544
+ processing capabilities is that the decision makers
545
+ may fail to remember critical pieces of data,
546
+ overlook stored information, draw hasty conclu-
547
+ sions, and produce flawed answers.Evidence of the effects of limitations in mem-ory and shared attention capacity on human deci-
548
+ sion making were found during baseline testing
549
+ (Hutchins & Kowalski, 1994; Hutchins & Westra,
550
+ 1995) and during empirical evaluation of the DSS
551
+ (Kelly, Hutchins, & Morrison, 1996; Kelly, Mor-
552
+ rison, & Hutchins. 1996).Simon (1978, p. 273) states, "...the human in-formation processing system...operates almost en-
553
+ tirely serially, one process at a time, rather than in
554
+ parallel fashion. This seriality is reflected in the
555
+ narrowness of its momentary focus of attention."
556
+ However, the AAW problem forces the decision
557
+ maker to operate in a parallel processing mode
558
+ when several contacts demand attention at the
559
+ same time. The requirement to monitor and
560
+ maintain an accurate SA for these concurrentcontacts, over the course of the evolving situation,imposes an additional load of strategically man-
561
+ aging the overall situation. Several researchers
562
+ have argued that "managing the attentional and
563
+ conceptual processes that permit cogent SA in-
564
+ volves significant cognitive resources" (Adams, et
565
+ al, 1995, p. 91; Endsley, 1988). The tasks of pri-
566
+ oritizing contacts and the associated actions to be
567
+ taken by the team, updating the status of critical
568
+ contacts, responding to the other requisite tasks in
569
+ the queue and, more generally, of strategically
570
+ managing the workload of current multitask sys-
571
+ tems under dynamically changing scenarios can
572
+ place an unrealistic cognitive load on the decision
573
+ maker.A major advantage offered by the experimen-tal DSS is that it should "buy time" for the user by
574
+ (1) performing many of the cognitive processing
575
+ tasks for the user and (2) by presenting informa-
576
+ tion in graphic format. The DSS will synthesize
577
+ much of the information used to develop situation
578
+ awareness and present a coherent picture of the
579
+ situation to the user. This integrated picture will
580
+ be portrayed graphically—rather than in the cur-
581
+ rent text-based format—which should further re-
582
+ duce the amount of time required to assimilate
583
+ this information. By performing several informa-
584
+ tion processing steps for the decision maker the
585
+ decision maker's limited cognitive resources can
586
+ be used for the types of decisions which require
587
+ human abilities (e.g., the decision on whether to
588
+ engage).5.1 Working Memory RequirementsAn essential information processing step re-quired by this task—and one which levies a heavy
589
+ load on working memory—involves integrating
590
+ kinematic and sensor variables and maintaining an
591
+ awareness of changes in these variables over time.
592
+ Changes in a contact's behavior such as, decreas-
593
+ ing altitude, increasing speed, changes in elec-
594
+ tronic emissions, etc., can provide key indicators
595
+ of possible hostile intent. With current systems,
596
+ the decision maker receives numerous reports
597
+ from CIC team members who provide various
598
+ pieces of the overall tactical picture (such as,
599
+ kinematic parameter values, active electronic
600
+ emitter identifications, and behavioral responses
601
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 10Monterey, Ca.of the contact in response to queries by the team)regarding a particular contact. Some of this infor-
602
+ mation is also displayed in a text-based format for
603
+ the user when a contact is "hooked" (that is, se-
604
+ lected for display) by the decision maker. How-
605
+ ever, to recognize a change in certain variables,
606
+ current systems require the user to retain parame-
607
+ ter values in short-term memory in order to recog-
608
+ nize a change in the parameter, such as altitude.When the decision maker is monitoring sev-eral concurrent contacts (such as, cycling through
609
+ three or four contacts in a 1-minute period) human
610
+ working memory capabilities may quickly be sur-
611
+ passed. To detect a change in a critical parameter
612
+ value, the decision maker must maintain the pa-
613
+ rameter values for the contacts of interest in
614
+ working memory as he or she cycles between sev-
615
+ eral contacts. For example, the decision maker
616
+ must be able to recall that contact 7022 was at
617
+ 14,000 ft. altitude one minute ago, and then sub-
618
+ tract the current altitude value of 10,000 ft., which
619
+ will then indicate the aircraft is in a rapid descent.
620
+ The DSS was developed to aid the decision maker
621
+ by performing several of these cognitive process-
622
+ ing tasks, thus, reducing the cognitive load for the
623
+ user. By presenting the synthesized picture of the
624
+ contact's behavior over time, through the use of
625
+ graphical displays, critical changes should be
626
+ immediately apparent to the user.A second memory-intensive task involvesmaintaining, in working memory, a current list of
627
+ actions taken by team members, the contact's re-
628
+ sponse to these actions taken by the CIC team,
629
+ and pending actions. Research has established that
630
+ "memory is limited and that list maintenance is
631
+ effortful and fallible—more so if the list must be
632
+ ordered and still more if the membership of the
633
+ list must be dynamically reordered and modified
634
+ during retention" (Bower, 1970, as cited in Ad-
635
+ ams, et al, 1995, p. 91). The DSS should reduce
636
+ the cognitive effort required for distributing atten-
637
+ tion among the many contacts to be attended to
638
+ and actions that are required. Working memory
639
+ requirements should be reduced by having the
640
+ DSS act as an intelligent "assistant," reminding
641
+ the user regarding what actions are to be taken and
642
+ when the actions are to be taken.A third way the DSS will reduce memory andinformation processing requirements is by dis-
643
+ playing templates depicting weapons' envelopes
644
+ for both the inbound contact and "own-ship." This
645
+ should facilitate critical comparisons and judg-
646
+ ments regarding timing of actions. During a sce-
647
+ nario decision makers have to either rely on mem-
648
+ ory to recall the launch range for various weapons
649
+ or query a team member for this information. Both
650
+ of these methods waste limited resources. The
651
+ high workload and high tempo characteristic of
652
+ littoral scenarios produce a stress-ful decision-
653
+ making environment. The phenom-enon that in-
654
+ creasing stress leads to decreasing working mem-
655
+ ory is well documented (e.g., Hockey, 1986). The
656
+ latter method for obtaining the desired informa-
657
+ tion wastes limited resources by increasing the
658
+ communications load and requiring more time to
659
+ wait for a team members' response to the queryUnder these high-tempo and high workloadconditions human memory and attentional re-
660
+ sources can easily be surpassed. Several cogni-
661
+ tively resource intensive information processing
662
+ steps are eliminated for the human decision maker
663
+ by having them performed by the DSS. We pre-
664
+ dict that the decision support tools will reduce the
665
+ cognitive workload imposed on the decision
666
+ maker in the following three ways: (1) by reduc-
667
+ ing the amount of information processing to be
668
+ performed, (2) reducing working memory re-
669
+ quirements, and (3) assisting the user in allocating
670
+ limited attentional resources.5.2 Reducing Human ErrorThe study of human cognitive processes andrelated error mechanisms has gained rapidly in-
671
+ creasing interest in the past decade. Rasmussen
672
+ (1987) argues that the emphasis in attempting to
673
+ understand human errors must shift from tasks to
674
+ the human-task mismatch. For example, Gruner
675
+ (p. 39), in discussing the Vincennes incident,
676
+ maintains that "the system was poorly suited for
677
+ use by human beings during rapid military action."
678
+ He ascribes this lack of suitability to a human-
679
+ machine mismatch between the rate of data flow
680
+ possible with modern computer systems that can
681
+ process and display information at phenomenal
682
+ data rates and the "comprehension capability of
683
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 11Monterey, Ca.users which has remained almost static for thou-sands of years." This causal approach to under-
684
+ standing human error is based on the premise that
685
+ errors are rarely random and can be traced to
686
+ causes and contributing factors. Once these con-
687
+ tributing factors are identified they can be miti-
688
+ gated.The impact and vulnerability of systems andhuman interfaces, because of incompatibilities
689
+ between the way people perceive, think, and act,
690
+ are documented in the popular and technical lit-
691
+ erature (Buck, 1989; Casey, 1993; Norman, 1988;
692
+ Perrow, 1984; Wilson & Zarakas, 1978). Newly
693
+ developed systems will succeed or fail based on
694
+ our ability to minimize these incompat-ibilities
695
+ between the characteristics of the things we create
696
+ and the way we use them. There are many well-
697
+ documented instances of critical systems or pa-
698
+ rameter changes going unnoticed or unheeded be-
699
+ cause the operating procedures, or the human ma-
700
+ chine interface, provided no historical trace. For
701
+ example, an unnoticed increase in altitude con-
702
+ tributed to the shoot down of the Iranian airbus by
703
+ a U.S. Navy ship—when the team mistakenly be-
704
+ lieved the aircraft to be descending—because
705
+ there was no historical trace to make the aircraft's
706
+ actual increasing altitude apparent (Dotterway,1992). Five personnel in the U.S.S. Vincennes's
707
+ combat information center, all viewing separate
708
+ displays, reported the aircraft as descending while
709
+ the Aegis data tapes later revealed a flight pattern
710
+ of ascent (Roberts & Dotterway, 1995). One of the
711
+ official investiga-tions of this incident, the Fo-
712
+ garty Report (1988, p. 45), states that "stress, task
713
+ fixation, and an unconscious distortion of data
714
+ may have played a major role in this incident." A
715
+ panel of five psychologists from the American
716
+ Psychological Association who testified before
717
+ Congress concluded that there were "predictable
718
+ failings of human judgment under intense stress
719
+ compounded by complex technology [which]
720
+ clearly contributed to the accidental shooting of
721
+ Iranian airliner Flight 655" (APA, p. 4).It is generally accepted that between 60-80percent of the accidents and malfunctions in
722
+ transportation, manufacturing, process control,
723
+ weapon, and other systems are attributable to hu-
724
+ man error (Senders & Moray, 1991; Van Cott,1993; Weiner, 1994). Reducing tactical decisionmaking errors is one goal of the TADMUS pro-
725
+ gram. The following section presents a brief re-
726
+ view of an experiment conducted to develop a
727
+ baseline on tactical decision making performance
728
+ in response to fairly stressful scenarios. A com-
729
+ panion paper (Hutchins, Kelly, & Morrison, 1996)
730
+ describes the experimental DSS modules and the
731
+ way they are hypothesized to enhance tactical de-
732
+ cision making performance.6 TADMUS Baseline ExperimentEarly research involved data collection in the De-cision-Making Evaluation Facility for Tactical
733
+ Teams (DEFTT) Laboratory using simulated ex-
734
+ isting shipboard displays to establish a baseline on
735
+ decision-making performance. The purpose of this
736
+ effort was to document baseline decision-making
737
+ performance for experienced naval officers. Dur-
738
+ ing the baseline phase of testing, a detailed under-
739
+ standing was developed of the cognitive processes
740
+ underlying the various tasks involved in situation
741
+ assessment—and where the bottlenecks occur.
742
+ This understanding was then used to design the
743
+ way the information is presented to the user in or-
744
+ der to facilitate performance of the required tasks.6.1 SubjectsThis study focused on the command-level de-cision makers of an antiair warfare team on an
745
+ Aegis cruiser—the commanding officer and the
746
+ tactical action officer. Subjects in the study con-
747
+ sisted of six commanding officer/tactical action
748
+ officer teams drawn from twelve active duty Na-
749
+ val personnel; some were from training com-
750
+ mands while others were from operational com-
751
+ mands aboard ship or assigned to group staffs.6.2 ProcedureData were collected in the DEFTT Labora-tory, a six-station test-bed environment that
752
+ simulates console positions in a Navy Aegis
753
+ cruiser combat information center. (For a detailed
754
+ description of the DEFTT Laboratory see Hutch-
755
+ ins, 1996.) Four stations were filled by confeder-
756
+ ates (active duty Navy personnel) who play antiair
757
+ warfare support-team member roles. These roles
758
+ included the antiair warfare coordinator, identifi-
759
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 12Monterey, Ca.cation supervisor, tactical information coordina-tor, and electronic warfare supervisor. After ap-
760
+ proximately 1 1/2 hours of orientation to the labo-
761
+ ratory and training in the use of the computer con-
762
+ soles the subjects engaged in four scenarios. The
763
+ scenarios were each about 25 minutes in length
764
+ and contained between 11 and 14 contacts of in-
765
+ terest per scenario, in addition to numerous back-
766
+ ground contacts.6.3 Treatment of DataTeam communications were recorded on amultichannel audio recorder; these included all
767
+ intra-team exchanges, as well as all communica-
768
+ tions with simulated off-ship personnel. Audio
769
+ tapes were used to produce verbatim, time-
770
+ stamped transcripts of all team communications.
771
+ A modified version of the TapRoot® Incident In-
772
+ vestigation System (Paradies, 1991; Paradies and
773
+ Unger, 1991) was then applied to identify errors.
774
+ The objective was to identify tactically significant
775
+ errors committed during the scenario. Tactically
776
+ significant errors were defined as those errors that
777
+ may lead to loss of life or significant political em-
778
+ barrassment. The following criteria were used for
779
+ counting an error as tactically significant: (1) loss
780
+ of situation awareness, (2) failure to take defen-
781
+ sive action when within the weapon's range of an
782
+ approaching contact, or (3) a violation of rules of
783
+ engagement (ROE). Video recordings were made
784
+ of the commanding officer and tactical action of-
785
+ ficer computer screens. Detailed analyses of all
786
+ audio and video recordings were conducted. (For
787
+ a more detailed coverage of the methodology and
788
+ results see Hutchins and Westra, in preparation).6.4 ResultsThe complex, time-constrained, decision-making situations embodied in the experimental
789
+ scenarios resulted in a large number of decision
790
+ errors. The mean number of tactically significant
791
+ errors documented across six teams and four sce-
792
+ narios was 14; the number of errors ranged from
793
+ nine to twenty-two. The standard deviation was
794
+ 3.7. Subjects performed an average of 50% of the
795
+ required behaviors as specified in the rules of en-
796
+ gagement. The ordinal agreement between three
797
+ raters (navy subject matter experts) on error countranks from TapRoot® analyses was computed.Results showed a high degree of agreement with
798
+ the Kendall's W of .93 indicating that 93% of the
799
+ possible rank variance is accounted for.6.4.1 Decision-Making ErrorsDetailed examinations of the informationprocessing sequences performed during tactical
800
+ decision making have revealed a variety of errors.
801
+ On average, subjects failed to take required ac-
802
+ tions, about half of the time. Explanations based
803
+ in the cognitive psychology literature have been
804
+ pursued, as a major goal of the TADMUS pro-
805
+ gram is to develop a DSS based on an under-
806
+ standing of the way in which human decision
807
+ makers actually process information under rapidly
808
+ evolving situations.The majority of documented errors involvederrors of omission, that is, "failure to take defen-
809
+ sive measures” and "failure to adhere to ROE.”
810
+ Failure to take defensive measures included fail-
811
+ ure to take actions to defend own-ship when an
812
+ approaching aircraft had reached its weapon's re-
813
+ lease range. An example involved a case where
814
+ two contacts were within the specified ROE limit,
815
+ yet no action had been taken. The types of actions
816
+ included in the “failure to adhere to ROE” cate-
817
+ gory include failure to take action regarding the
818
+ items listed and defined below: (a) issuing warn-ings is part of the usual identification process andinvolves three levels of warnings with increasing
819
+ levels of urgency; (b) establish friendly force cri-teria refers to establishing a plan with otherfriendly ships in the area to coordinate how they
820
+ will respond to potential threats; (c) changes inkinematics/ identification friend or foe—subjectsare expected to notice significant kinematic
821
+ changes and/or identification friend or foe pa-
822
+ rameter changes; and (d) other identification pro-cedures includes actions such as illuminating withfire control radar.The other major category of tactically signifi-cant error involved “loss of SA.” Loss of SA er-
823
+ rors were grouped under errors of commission and
824
+ errors of omission and then further categorized
825
+ into subgroups. Fifty-five percent of the loss of
826
+ SA errors involved taking the wrong action (error
827
+ of commission) while 45% of the errors involved
828
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 13Monterey, Ca.failing to take some required action (error ofomission). Error categories included under errors
829
+ of commission involved incorrectly engaging a
830
+ track (3%), incorrectly warning a track (29%),
831
+ other incorrect actions (16%), and incorrect re-
832
+ porting (7%). The two instances of incorrectly en-
833
+ gaging an aircraft, which were F-1 Mirage aircraft,
834
+ were considered errors because the decision
835
+ maker failed to take certain actions prior to en-
836
+ gaging—not necessarily because the aircraft
837
+ should not have been engaged. The actions that
838
+ the decision makers failed to take involved ascer-
839
+ tain-ing the identification of the aircraft for one
840
+ case and failure to warn and illuminate the aircraft
841
+ prior to engaging for the second case. Most in-
842
+ stances of incorrectly issuing warnings to the air-
843
+ craft involved issuing the warning when the air-
844
+ craft was within its territorial airspace (that is, in-
845
+ side the 12 nautical mile limit which is interna-
846
+ tionally recognized as under control of that nation)
847
+ or issuing a warning at a level different from what
848
+ was required. Other incorrect actions included il-
849
+ luminating the aircraft, “locking up” with radar, or
850
+ ordering the aircraft to divert when the aircraft
851
+ was still within its territorial airspace. Incorrect
852
+ reporting involved inaccurate reports on the status
853
+ of the tactical situation (such as, indicating to the
854
+ battle group commander that certain actions had
855
+ been taken when they had not, misidentification of
856
+ an aircraft, or omitting critical tracks from a re-
857
+ port).Errors of omission categorized under the “lossof SA” category included: (a) failure to identify or
858
+ attend to a contact; (b) failure to take action (e.g.,
859
+ to issue “hold-fire” when a contact turned out-
860
+ bound); (c) failure to recognize a threat (e.g.,
861
+ designating an aircraft as a non-threat because it
862
+ had passed its closet point-of-approach, yet it was
863
+ still within missile-launch range); (d) instances of
864
+ confusion or forgetting (e.g., forgetting or ignor-
865
+ ing critical data, forgetting whether or not it had
866
+ been warned, illuminated, or “locked-on,” or for-
867
+ getting the aircraft's response, or lack of response
868
+ to these actions, forgetting the status of a contact,
869
+ and confusing contacts); (e) misperception of data
870
+ (e.g., reporting a contact as turning outbound
871
+ when it is still inbound); (f) unclear communica-
872
+ tion (issuing vague orders regarding actions to betaken by team member, such as, failure to specifywhich weapon system is to be used or which con-
873
+ tact is to be engaged).6.4.2 Cognitive explanationsThe cause of failures to take required actionsis, in many cases, attributed to the extremely high
874
+ task demands levied on the decision maker by the
875
+ scenario and the human decision-maker's limited
876
+ attentional resources. Many cases are also attrib-
877
+ uted to working memory limitations. Maintaining
878
+ an awareness of the status of each contact and the
879
+ status of many actions to be taken by the antiair
880
+ warfare team—which actions have been taken and
881
+ what the contact's response to the action was—
882
+ severely taxes the decision maker's working
883
+ memory. The high workload entailed in the
884
+ scenarios produces a highly time-compressed
885
+ decision making situation. This time-compressed
886
+ decision making situation—where attentional
887
+ resources and working memory capacity are
888
+ limited—do not allow the decision maker to
889
+ maintain accurate SA for all tracks at any given
890
+ time. We anticipate that the decision support
891
+ modules in the DSS will mitigate these types of
892
+ errors.Human information processing capabilities arenot well suited to dealing with a "multiplicity of
893
+ simultaneous and disjointed tasks. Thoughtful at-
894
+ tention is modular: People can consciously think
895
+ about only one thing at a time" (Adams, et al,
896
+ 1995, p. 92). As a result, they do not handle inter-
897
+ ruptions very well. Research indicates that when
898
+ an operator is faced with as few as two tasks that
899
+ consist of merely the detection or recognition of
900
+ simple signals, a cost may be incurred in terms of
901
+ a significant loss in sensitivity or time that can be
902
+ allocated to either by the requirement to divide or
903
+ switch attention between them (Broadbent, 1957;
904
+ Schneider and Detweiler, 1988; Swets, 1984).The memory demands of managing complex,multi-task situations can easily surpass human
905
+ limitations. The decision maker must not forget
906
+ any of the contacts or tasks requiring action. In
907
+ addition to remembering all the tasks needing at-
908
+ tention, however, are the complexities entailed in
909
+ keeping track of the data and substeps associated
910
+ with each contact and prior action. The aviation
911
+ literature provides many examples of incidents
912
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 14Monterey, Ca.with explanations similar to the root causes forerrors that were found in the TADMUS program.
913
+ One category includes the potentially disastrous
914
+ effects of interruptions in the task for air traffic
915
+ controllers and pilots. Similarly, in the AAW en-
916
+ vironment, momentary intervening attention to
917
+ another task or contact, or an interruption in a
918
+ procedure can leave the procedure, or processing
919
+ of a contact incomplete with potentially cata-
920
+ strophic results.A fairly consistent pattern of tactical decision-making errors was documented from data col-
921
+ lected during the baseline data collection period.
922
+ The root causes of these errors were traced to
923
+ cognitive mechanisms such as limited attentional
924
+ resources and working memory limitations. By
925
+ developing an understanding of the pattern and
926
+ types of errors most frequently observed in this
927
+ task domain we hope to provide a DSS which will
928
+ mitigate these errors.7 DiscussionFailure to take appropriate actions may be ex-plained by the limited resource capacity of human
929
+ memory. In these scenarios a large number of
930
+ contacts are monitored for changes in any of sev-
931
+ eral key parameters. Three modules in the DSS
932
+ are hypothesized to assist with recognizing a
933
+ problem and taking the appropriate actions: track
934
+ history; response manager; and the track priority
935
+ list and alerts.Features offered by the DSS to address errorsattributed to limited attentional resources include
936
+ focusing attention on (1) high priority contacts
937
+ (i.e., track priority list and alerts), as well as on (2)
938
+ missing data (e.g., basis for assessment), and (3)
939
+ enabling the decision maker to use more data than
940
+ is typically used in current systems (e.g., track
941
+ history, comparison to norms). Current systems
942
+ require the user to retain previous contact data in
943
+ memory to compare with current values for criti-
944
+ cal parameters. Current systems also require the
945
+ user to rely on recall of vast amounts of informa-
946
+ tion from training and experience. Presenting all
947
+ known data on a contact in a synthesized way
948
+ should reduce working-memory requirements and
949
+ facilitate recognition. Additional potential per-
950
+ formance enhancement features, offered by theDSS, include displaying the complete kinematiccontact history, presenting graphic displays of lo-
951
+ cation and trends, highlighting missing data, pro-
952
+ viding alerts, and providing assessments of cur-
953
+ rent contact identity that go beyond what existing
954
+ systems currently present.Focusing the user's attention on trend andhistory data should decrease the cognitive work-
955
+ load imposed by these scenarios where many
956
+ contacts must be identified and responded to un-
957
+ der severe time constraints. Similarly, delineating
958
+ trend and history data can assist in the identifica-
959
+ tion of a contact where noticing changes in critical
960
+ parameters is essential. Presentation of trend and
961
+ history data, as well as threat assessment and
962
+ comparison to norms, should also mitigate cogni-
963
+ tive "tunnel vision" effects where the decision
964
+ maker attends to a smaller number of cues when
965
+ under stress.The notion of time is an important character-istic of situation awareness (Harwood, Barnett,
966
+ and Wickens, 1988). The past is critical to under-
967
+ standing the present, and both past and present
968
+ information must be used to predict future events
969
+ (Shrestha, et al, 1995). Endsley (1988) referred to
970
+ the "projection of their (perceived elements) status
971
+ in the near future" when discussing situational
972
+ awareness. However, Endsley also noted the task
973
+ of attending to incoming information and subse-
974
+ quently predicting future events places a heavy
975
+ load on working memory. Several decision sup-
976
+ port modules were developed to assist the user in
977
+ remaining aware of the contact's history and
978
+ changes over time. Remembering which actions
979
+ are to be taken at what time levies an additional
980
+ burden by placing a heavy load on working mem-
981
+ ory. A secondary time savings should be achieved
982
+ by the DSS acting as an intelligent "advisor," that
983
+ is, by assisting the decision maker in knowing
984
+ what actions to take, when to take them, and
985
+ which actions have already been taken. A tertiary
986
+ time savings can be achieved by including a tem-
987
+ plate depicting the weapons' release ranges so the
988
+ decision maker does not need to rely on fallible
989
+ human memory or query a team member regard-
990
+ ing weapons ranges. By graphically depicting a
991
+ synthesized view of a contact's kinematic history,
992
+ with the focus on changes in the contact's behav-
993
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 15Monterey, Ca.ior over time, along with the contact's weaponsenvelope in relation to both own-ship's radar and
994
+ weapon coverage, information processing time
995
+ can be saved for the decision maker.8 ConclusionsThe research reported here focuses on developinga DSS which reflects the natural decision-making
996
+ strategies of humans. Presenting synthesized in-
997
+ formation in the form of graphic presentations is
998
+ expected to reduce the cognitive processing load
999
+ for the decision maker when performing situation
1000
+ assessment. The intention is to aid the decision
1001
+ maker by providing information in a way that will
1002
+ minimize the need to maintain information in
1003
+ working memory, reduce information processing
1004
+ demands, help focus attentional resources on the
1005
+ highest priority contacts, remind the user of ac-
1006
+ tions which need to be taken, help make decisions
1007
+ under stress, and support higher levels of situation
1008
+ awareness.Decision support systems require that the hu-man's strengths be used in synergy with the ad-
1009
+ vantages offered by the DSS. Limitations associ-
1010
+ ated with the current generation of automated de-
1011
+ cision aids include the idea that (1) they cannot
1012
+ adequately capture the expertise developed by ex-
1013
+ perience over time and (2) since all contingencies
1014
+ cannot be anticipated, the expert's abilities to use
1015
+ intuition is indispensable (Mosier, in press). Mo-
1016
+ sier's review of the limitations of automated deci-
1017
+ sion systems delineates the characteristics of hu-
1018
+ man expertise that surpass the capabilities of
1019
+ automated systems. These include the human ca-
1020
+ pacity for creativity, adaptability, the ability to
1021
+ incorporate experience, the presence of a broad
1022
+ focus, analogical reasoning, and commonsense
1023
+ knowledge. The goal for the DSS is to capitalize
1024
+ on the strengths of the human along with the ad-
1025
+ vantages provided by the decision support system.9 Testing the DSSThe prototype DSS display modules are currentlybeing empirically evaluated in the simulated tacti-
1026
+ cal environment provided in the DEFTT Labora-
1027
+ tory. Experienced naval decision makers engage
1028
+ in experimental scenarios with and without access
1029
+ to the DSS display. The various decision supportmodules will be tested individually and in combi-nation in future experiments. Data on reduction of
1030
+ errors, improvements in users' situation awareness
1031
+ scores, changes in communication patterns, and
1032
+ subjective responses to the decision support sys-
1033
+ tem will be collected.10 Future ResearchWhile tools based on both the RPD and explana-tion-based reasoning models of decision making
1034
+ are included in the DSS there is no direct connec-
1035
+ tion between the two. Research is currently being
1036
+ conducted to extend schema theory to dynamic
1037
+ decision-making situations. This involves devel-
1038
+ oping and testing a hybrid model of cognitive be-
1039
+ havior in decision making to incorporate both
1040
+ types of knowledge, i.e., feature matching and
1041
+ story generation, as elements of the same schema
1042
+ model of naturalistic decision making (Smith &
1043
+ Marshall, in press). Schema theory as described
1044
+ by these authors, offers a context for integrating
1045
+ these two models which have typically been
1046
+ viewed as separate entities.AcknowledgmentsThe authors gratefully acknowledge the assistanceof Steve Francis, Brent Hardy, C.C. Johnson, Pat
1047
+ Kelly, Ron Moore, Connie O’Leary, Pat Marvel,
1048
+ Mike Quinn, and Will Rogers in data collection,
1049
+ interpretation, developing the DSS, and in con-
1050
+ ducting this research.ReferencesAdams, M. J., Tenney, Y. J., & Pew, R. W.(1995). Situation Awareness and the Cognitive
1051
+ Management of Complex Systems. HumanFactors, 37(1), 85-104.American Psychological Association, ScienceAgenda, Fall, 1988.Bower, G. H. (1970). Organizational Factors inMemory. Cognitive Psychology, 1, 18-46.Broadbent, D. E. (1957). A mechanical model forhuman attention and immediate memory. Psy-chological Review, 64, 205-215.Buck, L. (1989). Human error at sea. HumanFactors Bulletin, September, 32 (9), 12.
1052
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 16Monterey, Ca.Casey, S. (1993). Set Phasers on Stun and OtherTrue Tales of Design, Technology, and Hu-
1053
+ man Error. Aegean Publishing Company,Santa Barbara, CA.Casner, S. (1991). Task-Analytic Design ofGraphic Presentations. Technical Report AIP-145. Departments of Computer Sci-ence and Psychology, Carnegie Mellon Uni-
1054
+ versity and Learning Research and Devel-
1055
+ opment Center, University of Pittsburgh, PA.Cohen, M. S. (1993). The Bottom Line: Natu-ralistic Decision Aiding. In G. Klein, J. Ora-
1056
+ sanu, R. Calderwood, & C. E. Zsambok (Eds.)Decision Making in Action: Models andMethods (pp. 138-147). Ablex PublishingCorporation, New Jersey.Dawes, R. M. (1971). A case study of graduateadmissions: Application of three principles of
1057
+ human decision making. American Psycholo-gist. 34, 571-582.Dotterway, K. A. (1992). Systematic Analysis ofComplex Dynamic Systems: The case of the
1058
+ USS Vincennes. Unpublished master's thesis,Naval Postgraduate School, Monterey, CA.Endsley, M. R. (1988). Design and evaluation forsituational awareness enhancement. Pro-ceedings of the Human Factors 32nd Annual
1059
+ Meeting, 97-101.Federico, P. A. (1995). Expert and novice recog-nition of similar situations. Human Factors,37(1), 105-122.Formal Investigation into the Circumstances Sur-rounding the Downing of a Commercial Air-
1060
+ line by the USS Vincennes, by Rear Admiral
1061
+ William M. Fogerty, USN, 28 July 1988, pp.
1062
+ 4, 6. Hereafter referred to as the Fogerty Re-
1063
+ port.Gruner, W. P. No Time For Decision Making.(1990, November). U.S. Naval Institute Pro-ceedings, 39-41.Harwood, K., Barnett, B., and Wickens, C.(1988). Situational awareness: A conceptual
1064
+ and methodological framework. Proceedingsof the Symposium Psychology in the Depart-
1065
+ ment of Defense.Hockey, G. R. (1986). Changes in Operator Effi-ciency as a Function of Environmental Stress,
1066
+ Fatigue, and Circadian Rhythms. In: K. R.Boff, L. Kaufman, J. P. & Thomas (Eds.):Handbook of Perception and Human Per-formance. Wiley, New York.Hutchins, S. G. (in press). Decision-MakingEvaluation Facility for Tactical Teams. Naval
1067
+ Command, Control, and Ocean Surveillance
1068
+ Center, RDT&E Division Technical Report, in
1069
+ press, San Diego, CA.Hutchins, S. G., Kelly, R. T. and Morrison, J. G.(in press). Decision Support for Tactical Deci-
1070
+ sion Making Under Stress. Proceedings of theSecond International Symposium on Com-
1071
+ mand and Control Research and Technology.June 1996, Monterey, CA.Hutchins, S. G. and Kowalski, J. T. (1993). Tac-tical Decision Making Under Stress: Prelimi-
1072
+ nary Results and Lessons Learned. Proceed-ings of the 10th Annual Conference on Com-
1073
+ mand and Control Decision Aids. June 1993,Washington, D. C.Hutchins, S. G. and Rummel, B. K. (1995). ADecision Support System for Tactical Deci-
1074
+ sion Making Under Stress. Proceedings of the
1075
+ First International Conference on Command
1076
+ and Control Research and Technology. June
1077
+ 1995, Washington, D. C.Hutchins, S. G. and Westra, D. P. (1995). Patternsof Errors Shown by Experienced Navy Com-
1078
+ bat Information Center Teams. Proceedingsof the 39th Annual Meeting of the Human
1079
+ Factors and Ergonomics Society, San Diego,CA. October 1995.Hutchins, S. G. and Westra, D. P. (in preparation).TADMUS Baseline Experimental Results.
1080
+ Naval Command, Control, and Ocean Sur-
1081
+ veillance Center, RDT&E Division Technical
1082
+ Report, in preparation, San Diego, CA.Kaempf, G. L. and Militelo, L. G., (1992). Deci-sion Making in Emergencies, First Offshore
1083
+ Installation Management Conference: Emer-
1084
+ gency Command Responsibilities Collected
1085
+ Papers, Aberdeen, Scotland.Kaempf, G. L. , Wolf, S. and Miller, T. E. (1993).Decision Making in the Aegis Combat Infor-
1086
+ mation Center. In Proceedings of the Human
1087
+ Factors and Ergonomics Society 37th Annual
1088
+ Meeting (pp. 1107-1111). Santa Monica, CA:
1089
+ Human Factors and Ergonomics Society.
1090
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 17Monterey, Ca.Kelly, R. T. , Hutchins, S. G., and Morrison, J. G.(1996). Decision Processes and Team Com-
1091
+ munications with a Decision Support System.Proceedings of the Second International Sym-posium on Command and Control Research
1092
+ and Technology. June 1996, Monterey, CA.Kelly, R. T. , Morrison, J. G., and Hutchins, S.G., (in press). Impact of Naturalistic Decision
1093
+ Support on Tactical Situation Awareness.Proceedings of the 40th Annual Meeting of theHuman Factors and Ergonomics Society,Philadelphia, PA. September 1996.Kirshenbaum, S. S. (1992). Influence of Experi-ence on Information-Gathering Strategies.Journal of Applied Psychology, 77, 343-352.Klein, G. A. (1989). Recognition-Primed Decisions. In W. R. Rouse (Ed.) Advancesin Man-Machine Systems Research (pp. 47 - 92), Vol. 5. JAI Press, Inc.Klein, G. A. (1993). A Recognition-Primed Deci-sion (RPD) Model of Rapid Decision Making.
1094
+ In G. A. Klein, J. Orasanu, R. Calderwood, &
1095
+ C. E. Zsambok (Eds.) Decision Making inAction: Models and Methods (pp. 138-147).Ablex Publishing Corporation, New Jersey.Larkin, J. H. and Simon, H. A. (1987). Why adiagram is (sometimes) worth 10,000 words.Cognitive Science, 1, 65-99.Mosier, (in press). Myths associated with Auto-mated Decision Aids. In G. A. Klein, & C. E.
1096
+ Zsambok (Eds.) Advances in Naturalistic De-cision Making: Research and Applications,Hillsdale, NJ: Erlbaum.Mundy, C. E., Jr. (1994). Thunder and Light-ning: Joint Littoral Warfare, Joint ForceQuarterly, 4, Spring, 45-50.Noble, D. (1989). Application of theory of cog-nition to situation assessment. Vienna, VA:Engineering Research Associates.Noble, D. (1993). A Model to Support Devel-opment of Situation Assessment Aids. n G.
1097
+ A. Klein, J. Orasanu, R. Calderwood, & C. E.
1098
+ Zsambok (Eds.) Decision Making in Action:Models and Methods (pp. 287-305). AblexPublishing Corporation, New Jersey.Norman, D. A. (1988). The Psychology of Eve-ryday Things. Basic Books, Inc. New York.Norman, D. A. and Bobrow, D. G. (1975). Onthe data-limited and resources-limited proc-
1099
+ esses. Cognitive Psychology, 7, 44-64.Office of Naval Technology. (1992). FY 1993Program Plan for Tactical Decision-Making
1100
+ Under Stress, Arlington, VA: July 1992.Paradies, M. (1991). Root Cause Analysis andHuman Factors. Human Factors Society Bul-letin, 34(8), 1-4.Paradies, M. & Unger, L. (1991). TapRoot Inci-dent Investigation System Manual. Volumes1-7. System Improvements, Inc. Knoxville,
1101
+ TN.Pennington, N. & Hastie, R (1992). Explainingthe Evidence: Tests of the Story Model of De-
1102
+ cision Making. Journal of Personality and So-cial Psychology, Vol. No. 2, 189-206.Pennington, N. & Hastie, R (1993). A theory ofExplanation-Based Decision Making. In G. A.
1103
+ Klein, J. Orasanu, R. Calderwood, & C. E.
1104
+ Zsambok (Eds.) Decision Making in Action:Models and Methods (pp. 188-201). AblexPublishing Corporation, New Jersey.Perrow, C. (1984). Normal Accidents: Livingwith High Risk Technologies. New York, Ba-sic Books, Inc.Rasmussen, J. (1986). Information Processingand Human-Machine Interaction. In A. P.
1105
+ Sage (Ed.) Series Volume 12, North-Holland,
1106
+ Amsterdam.Roberts, N. C. & Dotterway, K. A. (1995). TheVincennes Incident: Another Player on theStage? Defense Analysis Vol 11, No. 1, pp.31-45.Salthouse, T. A. (1992). Cognition and Context.Science, 257, 982-983.Sarter, N. B. and Woods, D. D. (1991). Situationawareness: A critical but ill-defined phe-
1107
+ nomenon. The International Journal of Avia-tion Psychology, 1(1), 45-57.Schneider, W., and Detweiler, M. (1988). Therole of practice in dual-task performance:
1108
+ Toward workload modeling in a connectionist/
1109
+ control architecture. Human Factors, 30, 539-566.Senders, J. W. & Moray, N. P. (1991). HumanError: Cause, Prediction, and Reduction.Lawrence Erlbaum Associates, New Jersey.
1110
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 18Monterey, Ca.Shrestha, L. B., Prince, C., Baker, D. P. and Salas,E., (1995). Understanding Situation Aware-
1111
+ ness: Concepts, Methods, and Training. Hu-man/Technology Interaction in Complex Sys-
1112
+ tems. Vol 7, W. B. Rouse (Ed.). San Fran-cisco: JAI Press.Simon, H. A. (1978). Information-processingtheory of human problem solving. In W. K.
1113
+ Estes (Ed.), Handbook of Learning and Cog-nitive Processes, Vol 5, Human InformationProcessing, New York: Wiley.Smith, D. E. and Grossman, J. D. (1993). Under-standing and Aiding Decision Making in
1114
+ Time-Constrained and Ambiguous Situations.Unpublished manuscript.Smith, D. E. and Marshall, S. (in press). Apply-ing Hybrid Models of Cognition in Decision
1115
+ Aids. In G. A. Klein, & C. E. Zsambok (Eds.)Advances in Naturalistic Decision Making:Research and Applications, Hillsdale, NJ:Erlbaum.Swets, J. A. (1984). Mathematical models of at-tention. In R. Parasuraman and R. Davies
1116
+ (Eds.), Varieties of Attention (pp. 183-242).New York: Academic.Tolcott, M. A. (1991). Understanding and AidingMilitary Decisions. Office of Naval ResearchEuropean Office. 27th International Applied
1117
+ Psychology Symposium, Stockholm, Sweden,
1118
+ June 1991.Wilson, G. L. & Zarakas, P. (1978). Anatomy ofa blackout. IEEE Spectrum, February, 339-346.
1119
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 19Monterey, Ca.The requirement to interleave a mul-tiplicity of tasks—although not necessary
1120
+ an ongoing characteristic of shipboard
1121
+ scenarios—represents the type of situation
1122
+ where providing decision support may
1123
+ make the critical difference in the outcome
1124
+ for a scenario. For example, during the
1125
+ experimental scenarios the decision mak-
1126
+ ers may have to perform the following: monitor ship location (relative toother ships and objects in the vicinity) monitor and apply rules of en-gagement to all applicable tracks in the
1127
+ local operating area
1128
+ 
1129
+  receive and send radio messagesto the battle group commander and other operating units in the area
1130
+  monitor tracks on the Aegis dis-play system and maintain situation
1131
+ awareness for all contacts of
1132
+ interest monitor performance of actionstaken by team members to assess the situation
1133
+  monitor the tactical action offi-cer's/commanding officer's performance maintain communications withCIC team members regarding their assessment of tracks and vari-ous actions taken
1134
+ In broad perspective, although teammembers spend much of their time in rou-
1135
+ tine activities, a number of different atten-
1136
+ tionally demanding, knowledge-intensive,
1137
+ and procedurally complex tasks may de-
1138
+ mand attention at any moment. Each of
1139
+ these tasks is usually triggered by a
1140
+ stimulus event, such as a communication
1141
+ from a team member or an alert, and, in
1142
+ order to obtain proper interpretation, mayrequire additional information-seeking be-havior. The cognitive challenge of select-
1143
+ ing and interpreting information to main-
1144
+ tain and revise one's SA is inherently
1145
+ complex. (Jager, Tenney, and Pew, 1995
1146
+ HF) Problems arise when, in the dynamic
1147
+ and multidimensional environments of
1148
+ some littoral antiair warfare scenarios, the
1149
+ situation-critical data become more time-
1150
+ compressed or ambiguous than humans
1151
+ can handle within the inherent time con-
1152
+ straints of the evolving scenario.Resources are such things as processing effort,the various forms of memory capacity, and
1153
+ communications channels (Bobrow & Norman,
1154
+ 1975, in Rasmussen et al).Topics to be discussed include: (1) a descriptionof the difficult tasks identified for analysis; (2) the
1155
+ general methodological approach; (3) develop-
1156
+ ment of the performance measures and issues
1157
+ related to their development; (4) discussion of the
1158
+ ; and (5) discussion of the types of errors made by
1159
+ decision makers and interpretations for the cause
1160
+ of these errors based in the cognitive psychology
1161
+ literature.
1162
+ Paper presented to the Second International Command and Control Research and Technology Symposium,20Monterey, Ca.tactical operations require decision makingconditions of time pressure, stress, am-
1163
+ biguous, inaccurate and missing informa-
1164
+ tion, uncertain communications, and
1165
+ shifting conditions. These conditions make
1166
+ it difficult to perform careful analysis prior
1167
+ to making decisions. Traditionally, most
1168
+ decision research and decision support
1169
+ system development have focused on well-
1170
+ defined tasks in carefully controlled envi-
1171
+ ronments. Recent research has suggested
1172
+ that tactical decision makers use experi-
1173
+ ence to generate a likely course of action
1174
+ and then evaluate its feasibility using
1175
+ mental simulation.One way to prevent the same type of casualtyfrom being repeated is to thoroughly investigate
1176
+ and analyze root causes of actual mishaps, as well
1177
+ as data collected in a simulated tactical
1178
+ environment, and apply the findings in a concrete
1179
+ manner to improve tactical decision making.According to the recognition-primed decision-making (RPD) model, experienced decision
1180
+ makers can make rapid, high-quality decisions by
1181
+ associating a situation directly with the actions
1182
+ that normally work well in that kind of situation.
1183
+ Roberts, K. H., Stout, S. K., and Halpern, J. J.(1994) Decision Dynamics in Two High
1184
+ Reliability Military Organizations. Man-agement Science, Vol. 40, No. 5, May1994, 614-624.Tetlock, P. E., Accountability and the Persever-ance of First Impressions, Social Psycho-logical Quarterly, 26(1983), 285-292.Tetlock, P. E., Accountability: The NeglectedSocial Context of Judgment and Choice, in
1185
+ L. L. Cummings and B. M. Straw (Eds.),
1186
+ Research in Organizational Behavior, Vol.7, JAI Press, Greenwich, CT, 1985, 297-332.The comparison to norms tool is based on acognitive model of human information processing
1187
+ which uses a feature matching strategy. The
1188
+ model proposes a data-driven process. As such,
1189
+ the comparison to norms tool is a knowledge-
1190
+ based tool with the knowledge represented as
1191
+ templates. Each template is a linked timeline
1192
+ associating individual features, through a series of
1193
+ feature matches, with expected actions. This
1194
+ allows the tool to make an assessment of the
1195
+ situation presented to the decision maker. A
1196
+ related module is the response manager. This
1197
+ assessment consists of a categorization of the
1198
+ situation (e.g., “hostile aircraft attacking”) and
1199
+ presentation of an associated template. The
1200
+ response manager module shows a template for an
1201
+ assessment of the current situation. This is a
1202
+ timeline display for template features, or events.
1203
+ The response manager module also provide inputs
1204
+ to the Prioritized track list, Alerts, and Responses
1205
+ and Tripwires.SABER is a model of another cognitivestrategy employed in making decisions. This
1206
+ strategy is known as explanation-based reasoning,
1207
+ or story generation. In this approach, available
1208
+ data are assembled into explanatory structures,
1209
+ with one structure for each possible conclusion.
1210
+ Each of the explanations attempts to explain how
1211
+ every piece of data can be accounted for in
1212
+ support of each conclusion, even though some of
1213
+ the data items would naturally contradict reaching
1214
+ some conclusions. Contradictory data are
1215
+ explained through the of internal assumptions. It
1216
+ is assumed that there are a fixed number of pre-
1217
+ defined possible conclusions and each data item
1218
+ points directly to one of those possible conclu-
1219
+ sions.Once the explanations are constructed,SABER evaluates them to determine which seems
1220
+ most plausible. Plausibility is based on three
1221
+ Paper presented to the Second International Command and Control Research and Technology Symposium,21Monterey, Ca.criteria: simplicity, completeness, and impor-tance. SABER provide data products which relate
1222
+ data to explanatory hypotheses for the current
1223
+ situation. Hypotheses are presented in rank order,
1224
+ with evidence for and against each hypothesis and
1225
+ missing data is presented below the corresponding
1226
+ hypothesis. For a more detailed description of the
1227
+ DSS the reader is referred to Hair & Pickslay,
1228
+ 1992; Hair, Pickslay, & Chow, 1992; Hutchins
1229
+ and Rummel, 1994.Hair, D. C. and Pickslay, K. (1992). Explanation-Based Reasoning in Decision Support
1230
+ Systems, Proceedings of the 9th AnnualConference on Command and Control De-cision Aids, June 1992, Monterey, CA.Hair, D. C., Pickslay, K. & Chow, S. (1992)Explanation-Based Decision Support in
1231
+ Real Time Situations. Proceedings of the1992 IEEE International Conference onTools with AI. Nov. 1992, Arlington, VA.Various studies have indicated that as much as 90
1232
+ percent of industrial and system failures are
1233
+ produced by human error (Senders & Moray,
1234
+ 1991).
1235
+ At the same time, a shift was occurring in U.S.
1236
+ Navy doctrine, away from a "blue-water" strategy
1237
+ to a doctrine of littoral operations aimed at
1238
+ potentially hostile regional powers.Tactical decision makers in today's oper-ating environment are required to perform
1239
+ complex tasks in a highly dynamic envi-
1240
+ ronment. Numerous interactive surface
1241
+ units and aircraft whose parameters are in
1242
+ flux and which must be continually
1243
+ sensed, processed, their future status pro-
1244
+ jected (development of hypotheses re-
1245
+ garding their future behavior) and actions
1246
+ taken to assure the successful outcome.Operating close to land presents additionalchallenges to the tactical decision maker.
1247
+ In littoral (i.e., near-land) settings, which
1248
+ most likely represent the majority of future
1249
+ anticipated naval conflicts, the decisions to
1250
+ be made are even more complex than they
1251
+ would be in full-scale warfare. When ci-
1252
+ vilian and neutral nation resources are in
1253
+ the conflict area incoming information
1254
+ carries an added element of uncertainty. In
1255
+ these situations, interpretation of the rules
1256
+ of engagement, contact identification, de-
1257
+ termining the capability and possible in-
1258
+ tent of the potential threat, and the
1259
+ shoot/no-shoot decision often pose ex-
1260
+ tremely difficult decision problems. (Since
1261
+ 70 percent of the world's population lives
1262
+ within 200 miles of the sea, most future
1263
+ contingencies are likely to involve littoral
1264
+ warfare (Mundy, 1994).)as opposed to replacing "the user's approach tothe problem" (emphasis in original, Cohen, 1993)with tools based on decision analysis or
1265
+ mathematical optimization,
1266
+ A considerably smaller number are attributable to
1267
+ other causes such as mechanical, electrical and
1268
+ materials failure (Meshkati, 1993). The purpose for this phase of the TADMUSprogram is to empirically evaluate the effective-
1269
+ ness of a DSS based on these recent approaches to
1270
+ decision support, to the extent possible,"This focus on why errors occuris...different from...the typical study of
1271
+ human errors which solely emphasize what
1272
+ occurs, a point of view which has received
1273
+ considerable criticism." (Rouse & Rouse,
1274
+ 1983, p. 539)
1275
+ Paper presented to the Second International Command and Control Research and Technology Symposium,22Monterey, Ca. DSS which will mitigate typical types oferrors. The objective during this phase of
1276
+ the research was to develop an under-
1277
+ standing of the decision-making problems
1278
+ presented by current and future Navy sce-
1279
+ narios in order to identify the types and
1280
+ forms of information that are likely to fa-
1281
+ cilitate performance of these activities.
1282
+ Rouse, W. B. & Valusek, J. (1993). Evo-lutionary Design of Systems to Support
1283
+ Decision Making. In G. A. Klein, J.
1284
+ Orasanu, R. Calderwood, & C. E.
1285
+ Zsambok (Eds.) Decision Making inAction: Models and Methods (pp.270-286). Ablex Publishing Corpora-
1286
+ tion, New Jersey.The Naturalistic Decision Making (NDM)model (Klein, 1989, 1993) seems more
1287
+ applicable than traditional decision-
1288
+ making models to the types of decisions
1289
+ involved in tactical decision-making.
1290
+ Early work conducted under the TADMUS
1291
+ program to determine the cognitive strate-
1292
+ gies employed by Navy AAW decision-
1293
+ makers found that when performing situa-
1294
+ tion assessment 87% of the time a recog-
1295
+ nitional strategy was used and 13% of the
1296
+ time story generation was used (Kaempf,
1297
+ Wolf, and Miller, 1993). We have applied
1298
+ these new models of human decision
1299
+ making—which parallel the cognitive
1300
+ strategies used by domain experts—to the
1301
+ design of a DSS for enhancing antiair war-
1302
+ fare tactical decision making. These two
1303
+ models for situation assessment are feature
1304
+ matching and story generation.
1305
+ (The scenarios were intentionally devel-
1306
+ oped to have many tracks with a high de-
1307
+ gree of uncertainty associated where it is
1308
+ not always clear whether a particular track
1309
+ should be engaged. Our interest in theTADMUS program was in gaining insightinto and aiding the decision process. The
1310
+ reader is referred to Hutchins, 1995, for a
1311
+ detailed coverage of performance meas-
1312
+ urement issues and scenario development
1313
+ issues.)
1314
+ These conditions include dynamic, fluid
1315
+ situations, time pressure, high risk, multi-
1316
+ ple decision makers, shifting and compet-
1317
+ ing goals, action-feedback loops, and
1318
+ situations with uncertain and incomplete
1319
+ data (Orasanu & Connolly, 1993).
AD1184872-1-91 ADDED
The diff for this file is too large to render. See raw diff
 
ADA242766-1-18 ADDED
File without changes
ADA262560-1-58 ADDED
File without changes
ARN34403-1-110 ADDED
The diff for this file is too large to render. See raw diff
 
ARN38160-FM_3-90-000-WEB-1.-1-480 ADDED
The diff for this file is too large to render. See raw diff
 
BCT-CDR-Tactical-Decision-Exercises-1st-Edition-SEP2018-1-88 ADDED
The diff for this file is too large to render. See raw diff
 
DOD-LAW-OF-WAR-1-1254 ADDED
The diff for this file is too large to render. See raw diff
 
NavigationRulesStandardSize-1-33 ADDED
The diff for this file is too large to render. See raw diff
 
PRINs_C2-1-22 ADDED
@@ -0,0 +1,1319 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 2Monterey, Ca.tact identification, intent, available responses andtheir consequences. For example, the close prox-
2
+ imity of U.S. Navy forces and potential adversary
3
+ forces makes interpreting the actions of an in-
4
+ bound aircraft who does not respond to radio
5
+ warnings much more difficult. Should the air-
6
+ craft's behavior be interpreted as an attack profile,
7
+ or does the pilot merely intend to harass, or does
8
+ the aircraft in question not carry the equipment
9
+ necessary to receive verbal warnings, leaving the
10
+ pilot unable to receive radio warnings directed
11
+ toward him and unaware of his precarious posi-
12
+ tion? In extreme cases there is no clear cut right or
13
+ wrong answer about a decision. Rapidly unfolding
14
+ events result in severe time pressure and severe
15
+ (often catastro-phic) consequences for errors.
16
+ While current real-time battle management sys-
17
+ tems are well-suited to the demands of all-out
18
+ conflicts, they may not be optimized for littoral
19
+ situations where human intervention in decision-
20
+ making is even more important (Office of Naval
21
+ Technology [ONT], 1992). (Since 70 percent of
22
+ the world's population lives within 200 miles of
23
+ the sea, most future contingencies are likely to
24
+ involve littoral warfare (Mundy, 1994).)Two unfortunate and highly publicized eventsfocused attention on the difficult types of deci-
25
+ sions confronting naval commanders and provided
26
+ the impetus for this research. In the case of the
27
+ U.S.S. Stark, the commander made the decision to
28
+ not engage an inbound aircraft which was be-
29
+ lieved to not be a threat to his ship, and 27 U.S.
30
+ naval personnel lost their lives as a result. In the
31
+ case of the U.S.S. Vincennes, the commander
32
+ made the decision to engage the inbound aircraft
33
+ believing it was a threat to his ship—which turned
34
+ out to be a commercial airliner—and all personnel
35
+ aboard the airliner were killed as a result. In rec-
36
+ ognition of the complex and difficult decisions
37
+ required in these types of situations the Tactical
38
+ Decision Making Under Stress (TADMUS) pro-
39
+ gram was initiated to conduct research in the areas
40
+ of human factors and training technology. The
41
+ objective is to develop and apply principles that
42
+ can help avoid these types of situations in the fu-
43
+ ture. This paper, and a two companion papers
44
+ (Hutchins, Kelly, & Morrison, 1996; Kelly,
45
+ Hutchins, & Morrison, 1996), report on a multi-year, multi-experiment research effort conductedunder the TADMUS program to apply recent de-
46
+ velopments in decision theory and human-system
47
+ interaction technology to the design of a decision
48
+ support system (DSS) for enhancing tactical deci-
49
+ sion making under highly complex conditions.1.1 "Naturalistic" and Classical Decision-Making Para-digmsIn the same time frame that these tragic acci-dents occurred, a radical shift was occurring in the
50
+ way psychologists viewed human decision mak-
51
+ ing. Research was now focused on experienced
52
+ decision makers performing their normal tasks in
53
+ natural settings. "Naturalistic" decision-making
54
+ research studies the decision strategies people ac-
55
+ tually use in bringing their expertise to bear under
56
+ challenging real-world conditions. Decision
57
+ making milieus encompassed under the naturalis-
58
+ tic paradigm include hospital emergency rooms,
59
+ aircrew flight coordination, military command and
60
+ control settings, process control systems, and po-
61
+ lice and fire units.The naturalistic perspective, also known as"everyday cognition," is based on a belief that
62
+ cognitive functions "elicited in natural settings
63
+ (are) likely to differ, either quantitatively, or
64
+ qualitatively, from those that occur in artificial or
65
+ contrived situations, and results from sterile and
66
+ contrived situations may not generalize to less
67
+ constrained and more natural environments"
68
+ (Salthouse, 1992, p. 982). There is a growing
69
+ body of work that demonstrates that experienced,
70
+ real-world decision makers rarely use traditional
71
+ resource intensive strategies to make decisions in
72
+ the face of dynamic, adverse conditions and time-
73
+ pressure (Kaempf & Militelo, 1992; Klein, 1989;
74
+ 1993). Instead, experts rely on their abilities to
75
+ recognize and appropriately classify situations:
76
+ these abilities are based on having much experi-
77
+ ence in the task domain. Once they know what
78
+ they are facing they also tend to know what re-
79
+ sponse option to apply, based on retrieval from
80
+ memory of typical responses and outcomes that
81
+ worked well in past similar situations. They use
82
+ the limited time available to evaluate the feasibil-
83
+ ity of that option before implementing it. Experi-
84
+ enced decision makers recognize the situation or
85
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 3Monterey, Ca.scenario based on a comparison of the features ofthe current situation with stored memory represen-
86
+ tations, or schemata. Schemata are highly inter-
87
+ con-nected clusters of knowledge concerning cer-
88
+ tain situations, or particular problem types, and
89
+ associated actions or solution procedures (Fede-
90
+ rico, 1995). Once the situation is recognized, so-
91
+ lutions are stimulated by activation of these mem-
92
+ ory representations.In contrast to the naturalistic perspective, ear-lier analytical methods applied in decision support
93
+ systems were primarily used for option generation
94
+ and evaluation, rather than for situation assess-
95
+ ment. Traditional decision theorists argue that op-
96
+ timum decision making involves thorough analy-
97
+ sis of all the available data and the evaluation of
98
+ all possible hypotheses; these approaches tend to
99
+ rely on extensive calculations designed to arrive at
100
+ optimal solutions. Processes for making decisions
101
+ where the person weighs the pros and cons of
102
+ various options and selects the one option that
103
+ provides the most benefit are described in the de-
104
+ cision making literature. The extensive time re-
105
+ quirements and complicated mathematical calcu-
106
+ lations involved (e.g., Multi-Attribute Utility
107
+ Analysis), however, make these approaches unre-
108
+ alistic for situations requiring rapid decision
109
+ making. Generally, these analytical strategies in-
110
+ volve the following steps (Kaempf & Militello,
111
+ 1992): specify all relevant features of the task;
112
+  identify the full range of options;
113
+  identify the key evaluation dimensions;
114
+  identify weights for each dimension;
115
+  rate each option on each dimension;
116
+  tabulate the results, and
117
+  select the best option.These analytical strategies may be appropriate for
118
+ inexperienced subjects making decisions about
119
+ novel tasks, but not for experienced personnel
120
+ making real-time decisions. In natural settings,
121
+ time constraints and the difficulty in assigning
122
+ weights and rating the dimensions involved render
123
+ classical analysis techniques untenable. Typically
124
+ in realistic settings experts employ recognition-
125
+ based reasoning, not classical analytical ap-
126
+ proaches. Experienced decision makers use their
127
+ extensive knowledge to seek information, identifyand interpret the problem, understand the signifi-cance, derive the intention (where possible),
128
+ model the situation (as time allows), select the
129
+ action, evaluate the choice, and anticipate the con-
130
+ sequence. "This decision cycle is distinctly differ-
131
+ ent from classical models, which are based on the
132
+ assumption that all options, outcomes, and prefer-
133
+ ences are known and calculated in advance"
134
+ (Federico, 1995, p. 106).Traditionally, research in decision making hasbeen directed largely toward situations in which
135
+ (1) decision makers have sufficient time to gener-
136
+ ate options, conduct option assessment, and select
137
+ a course of action; (2) the consequences of an in-
138
+ correct response are not immediately severe; (3)
139
+ decisions are reached via consensus of a group;
140
+ and (4) workload is manageable. Little research
141
+ has been conducted into the development of tacti-
142
+ cal decision support systems for use in naturalistic
143
+ situations characterized by time pressure, high
144
+ risk, uncertainty and information ambiguity, high
145
+ workload, team coordination demands and task
146
+ complexity (ONT, 1992).The central hypothesis for the research re-ported here is that presenting decision makers
147
+ with decision support tools which were designed
148
+ to parallel the cognitive strategies employed by
149
+ experts, as observed in naturalistic settings, will
150
+ reduce the number of decision making errors. This
151
+ is accomplished by developing the architecture
152
+ and algorithms to process information the same
153
+ way research indicates humans do under similar
154
+ circumstances.2 Tactical Decision Making TasksThe global tactical decision-making task involvesidentification of and responding to numerous
155
+ contacts. When an aircraft (or a surface contact) is
156
+ detected the CIC personnel work as a team to de-
157
+ termine the identity and to try to determine
158
+ whether or not the aircraft poses a threat. The high
159
+ degree of inherent ambiguity associated with
160
+ contact information can often make threat assess-
161
+ ment a very difficult task. This is because many
162
+ pieces of data fit multiple hypotheses regarding
163
+ threat assessment. The global response choices
164
+ (that is, engage, monitor, do nothing) are largely
165
+ determined by the ship's orders and the current
166
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 4Monterey, Ca.geopolitical situation. Specific actions (such as,change course, issue verbal warnings, illuminate
167
+ with radar, challenge with other sensors, etc.) de-
168
+ pend on the local conditions and the relative posi-
169
+ tions of the inbound contact of interest and own-
170
+ ship. Determining which of these actions is likely
171
+ to be effective depends on maintaining an accurate
172
+ threat assessment which requires "continual up-
173
+ dating in accordance with recurrent situation as-
174
+ sessments" (Sarter & Woods, 1991, p. 52).This decision problem presents a highly chal-lenging cognitive task, that is, making inferences
175
+ and deductions from incomplete and uncertain
176
+ information derived from multiple sources and
177
+ relating to several concurrent threats (or potential
178
+ threats) under time-compressed conditions. The
179
+ cognitive functions performed by the tactical deci-
180
+ sion maker are both data and resource limited
181
+ (Norman & Bobrow, 1975). Decisions are re-
182
+ source limited by the mental resources of the deci-
183
+ sion makers, who must maintain large amounts of
184
+ information in memory under conditions of high
185
+ workload and stress. The decisions are data lim-
186
+ ited by the inability of the sensors to provide
187
+ complete, error-free, unambiguous data to support
188
+ the identification process. In particular, the ex-
189
+ perimental scenarios were designed to follow the
190
+ pattern of being set in an ambiguous situation
191
+ where one or more threats of uncertain origin and
192
+ uncertain intent approach either own ship or the
193
+ ship being protected and may not respond to
194
+ warnings. Scenarios were designed to be highly
195
+ ambiguous, as this quality of uncertainty is in-
196
+ dicative of the types of decisions to be made in
197
+ current and future scenarios.2.1 Threat AssessmentIn the antiair warfare problem, threat assess-ment is particularly difficult because the available
198
+ information is often incomplete or ambiguous.
199
+ The ambiguity could be due to (a) the information
200
+ transmission characteristics of the transmission
201
+ medium, such as a radar transmission or a radio
202
+ report that is only intercepted on an intermittent
203
+ basis, (b) deliberate deceptive actions (such as ra-
204
+ dar jamming) by the pilot flying the aircraft, or (c)
205
+ the overlapping classification categories typical of
206
+ many parameter measure-ments. For example, air-craft can typically fly at altitudes ranging between2,000 and 40,000 feet. Generally, an aircraft that
207
+ is flying above 20,000 feet is not considered to be
208
+ a threat. Conversely, an aircraft below 10,000 ft is
209
+ considered to be more of a potential threat. How-
210
+ ever, aircraft flying in the middle range, (that is,
211
+ below 20,000 ft. and above 10,000 ft.) can be
212
+ much more difficult to categorize. Because many
213
+ aircraft do fly in this middle range, other variables
214
+ need to be considered in conjunction with altitude.
215
+ This same situation of overlapping categorization
216
+ categories exists for several other variables. These
217
+ variables include radars that are found on both
218
+ threat and non-threat platforms, country of origin,
219
+ and measures of course and speed. In the case of
220
+ speed, for example, when an aircraft flying at a
221
+ low altitude decreases speed this could be viewed
222
+ as indicative of a threat action (that is, slowing
223
+ down in order to obtain better targeting informa-
224
+ tion); however, at the same time, there could be
225
+ other viable explanations for an aircraft's de-
226
+ creasing speed.If the decision maker had access to all dataabout a contact approximately twelve variables
227
+ would be used to determine identity and to infer
228
+ intent. Two or three of these items, alone, do not
229
+ provide definitive answers because, in many
230
+ cases, these parameter values do not fall within
231
+ clear-cut ranges for a particular assessment cate-
232
+ gory (i.e., threat, non-threat). Thus, a single time
233
+ slice of information provides an incomplete pic-
234
+ ture of the situation. In the dynamic, ambiguous
235
+ conditions characteristic of littoral operations, the
236
+ rate and direction of change (data history) can
237
+ help one better assess the threat and predict the
238
+ future state of the situation. When the incoming
239
+ information changes over time, the integration of
240
+ information as it changes can help the user extract
241
+ the message (Kirshenbaum, 1992). The DSS was
242
+ designed to do precisely this: to facilitate the inte-
243
+ gration process and present a synthesized picture
244
+ of the situation to the user in a format that can be
245
+ quickly assimilated. The variables, that are used to
246
+ develop a threat assessment, can be divided into
247
+ two classes: sensor information (raw or computer-
248
+ processed information) and the contact's response,
249
+ or lack of response, to actions taken by the team.
250
+ These other actions, and the integration of the in-
251
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 5Monterey, Ca.formation received via the contact's response orlack of response to them, are necessary to clarify
252
+ the tactical picture.2.2 Situation AwarenessWhile recent increasing interest among re-searchers regarding the concept of situation
253
+ awareness (SA) has generated a debate on the pre-
254
+ cise definition of this term most researchers ac-
255
+ knowledge the importance of the concept. In gen-
256
+ eral, SA refers to the decision maker's moment-
257
+ by-moment ability to monitor and understand the
258
+ state of the complex system and its environment
259
+ (Adams, Tenney, & Pew, 1995). These authors
260
+ state the essential idea which is that when emer-
261
+ gencies arise, the completeness and accuracy of
262
+ the decision maker's SA are critical to the ability
263
+ to make decisions, revise plans, and manage the
264
+ system. Specific decision-making tasks included
265
+ under SA include the ability to: (1) maintain an
266
+ accurate perception of the surrounding environ-
267
+ ment (both internal and external to the ship); (2)
268
+ identify problems and/or potential problems; (3)
269
+ recognize a need for action; (4) note deviations in
270
+ the mission; and (5) maintain awareness of tasks
271
+ performed (Shrestha, Prince, Baker and Salas,
272
+ 1995). To maintain an accurate SA the decision
273
+ maker should take into account both information
274
+ that is available and that which can be activated
275
+ from memory (Sarter and Woods, 1991).However, a difficulty arises as a result of theheavy workload imposed by this process. When
276
+ the decision maker is faced with several concur-
277
+ rent contacts of interest, all of which have numer-
278
+ ous associated data items (i.e., as many as a
279
+ dozen), some or all of which may change over the
280
+ course of the scenario (e.g., intelligence, active
281
+ radar emitters, various kinematic parameters, etc.)
282
+ memory load easily exceeds human capacity.
283
+ Moreover, changing parameters may impart dif-
284
+ ferent interpretations to what is occurring. In some
285
+ cases the moment-to-moment attentional demands
286
+ of a tactical situation are relentless and unforgiv-
287
+ ing (such as, a terrorist aircraft directly inbound
288
+ toward "own-ship" which can result in "task fixa-
289
+ tion"), sometimes relevant background knowledge
290
+ is unavoidably incomplete (such as, an unfamiliar
291
+ aircraft), and sometimes the decision maker is al-ready thinking and working as hard as possible,even when there are no unanticipated events when
292
+ there is a high contact density (Adams, et al,
293
+ 1995). These instances provide a few illustrations
294
+ of situations that can degrade situation awareness.Complex information gathering and process-ing systems have been designed to aid the deci-
295
+ sion-maker in the past. However, these systems
296
+ often increase the decision-maker’s burden due to
297
+ the inherent system complexity and the failure to
298
+ design them in a way that they will fit the user's
299
+ cognitive processing limitations. Often, these
300
+ systems require operators to perform difficult
301
+ cognitive tasks under heavy workloads. They must
302
+ perceive, synthesize and determine the relevance
303
+ of a continual stream of incoming information,
304
+ often pertaining to several concurrent contacts,
305
+ while projecting future anticipated events and
306
+ making decisions regarding actions to be taken.
307
+ Decision makers must assess, compare, and re-
308
+ solve conflicting information, while making diffi-
309
+ cult judgments, and remembering the status of
310
+ critical contacts along with the contact's response
311
+ to actions taken by the CIC team. These decision-
312
+ making tasks are interleaved with other required
313
+ tasks, such as keeping other team members in-
314
+ formed (both on and off the ship). Furthermore,
315
+ these complex tasks are performed under condi-
316
+ tions where adverse environmental (noise, vibra-
317
+ tion, temperature extremes, etc.) and internal
318
+ stressors (boredom, fatigue, anxiety, and fear) are
319
+ part of the environment.3 Decision Support PrinciplesA case has been made that previous generations ofdecision support systems, which focus primarily
320
+ on solution optimization and base decision sup-
321
+ port on normative models of human decision
322
+ making, are less applicable than a DSS that par-
323
+ allels the cognitive strategies used by domain ex-
324
+ perts in situations characterized by time-
325
+ constrained situations with uncertain and ambigu-
326
+ ous data (Smith & Grossman, 1993). These
327
+ authors point out that rarely, if ever, were earlier
328
+ tactical decision aids intended as psychological
329
+ models of human cognitive behavior. Instead,
330
+ these aids performed "complex and burdensome
331
+ calculations, reducing the work-load on personnel,
332
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 6Monterey, Ca.speeding up the dissemination of information, andproviding more time for command decision mak-
333
+ ing" (Tolcott, 1991, p. 44).3.1 Feature Matching and Story GenerationWe have applied two of these new models ofhuman decision making—which parallel the cog-
334
+ nitive strategies used by domain experts—to the
335
+ design of a DSS for enhancing antiair warfare
336
+ tactical decision making. These two models that
337
+ people use in assessing a situation are feature
338
+ matching and story generation. The feature
339
+ matching model, described by Noble (1989), in-
340
+ volves an organization of memory, or "schemas,"
341
+ and information-processing where decision mak-
342
+ ers use their previous experiences to assess a
343
+ situation and identify promising actions. Incoming
344
+ information is categorized, selected, edited, and
345
+ organized on the basis of a person's general
346
+ knowledge about a domain. Both story generation
347
+ and feature matching occur under conditions
348
+ where a large base of implication-rich, condition-
349
+ ally dependent pieces of evidence must be evalu-
350
+ ated before choosing an alternative from a set of
351
+ prospective courses of action. The feature match-
352
+ ing model applies a spatio-temporal dependence,
353
+ whereas story generation is an example of causal
354
+ dependence. According to the explanation-based
355
+ model, decision makers construct a causal model
356
+ to explain the available evidence (Pennington &
357
+ Hastie, 1993). At the same time, the decision
358
+ maker creates a set of alternatives from which an
359
+ action will be chosen. A decision is made when a
360
+ story is successfully matched to an alternative in
361
+ the choice set. Story generation occurs in complex
362
+ situations where the decision maker may not have
363
+ all the necessary information or when a series of
364
+ facts may appear to contradict each other. The de-
365
+ cision maker must then develop causal links be-
366
+ tween these facts to produce a coherent picture of
367
+ the situation (Klein, 1989; 1993).The explanation-based reasoning model isbased on research which found that jurors develop
368
+ a narrative story to organize trial information
369
+ where causal and intentional relations between
370
+ events are central (Pennington & Hastie, 1992).
371
+ Pennington & Hastie propose four certainty prin-
372
+ ciples—coverage, coherence, uniqueness, andgoodness-of-fit—that govern (i) which story willbe accepted, (ii) which decision will be selected,and the (iii) confidence or degree of certainty withwhich a particular decision will be made. This or-
373
+ ganization of the evidence by the decision maker
374
+ is believed to facilitate evidence comprehension.
375
+ A central component of this model is that the
376
+ story the juror constructs determines the juror's
377
+ decision. This explanation-based decision process
378
+ is employed when the body of evidence relevant
379
+ to a decision is large, complex, and the implica-
380
+ tions of its components are interdependent.Feature matching, also referred to as the rec-ognition-primed decision (RPD) model, "occurs
381
+ when the decision maker recognizes the features
382
+ of the present situation as similar or identical to
383
+ those of a previous situation" (Kaempf & Militelo,
384
+ 1992, p. 6). An adequate match triggers recall of
385
+ information learned about this type of situation:
386
+ (a) plausible goals, (b) critical cues to be moni-
387
+ tored, (c) expectations of what should happen, and
388
+ (d) a course of action that worked in similar situa-
389
+ tions. According to this recent approach, expert
390
+ decision makers may rely on well-developed
391
+ memory representations to guide decision making
392
+ in new (but similar) situations. The RPD model of
393
+ decision making fuses two processes—situation
394
+ assessment and mental simulation (Klein, 1993).
395
+ In the simplest case the situation is recognized as
396
+ familiar or prototypical, using feature matching,
397
+ and the obvious response is implemented. In a
398
+ more complex case the decision maker performs a
399
+ conscious evaluation of the response, using men-
400
+ tal simulation to uncover problems prior to im-
401
+ plementing the response. In the most complex
402
+ case the evaluation reveals flaws requiring modi-
403
+ fication, or the option is judged inadequate and
404
+ rejected in favor of the next most typical reaction.3.2 Situation AssessmentIn general, the overall task of responding toantiair warfare scenarios consists of situation as-
405
+ sessment ("what's going on") and course of action
406
+ selection ("what to do about it"). Recent theories
407
+ of decision making emphasize the importance of
408
+ situation assessment for good decision making in
409
+ naturalistic, event-driven situations. Moreover,
410
+ they stress that decisions regarding actions to be
411
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 7Monterey, Ca.taken are a by-product of developing the situationawareness that precedes action selection. Klein
412
+ (1989) has found that usually the situation itself
413
+ either determines or constrains the response op-
414
+ tions and that experienced decision makers make
415
+ up to 90% of all decisions without considering
416
+ alternatives. If the situation appears similar to one
417
+ that the decision maker has previously experi-
418
+ enced, the pattern will be recognized and the
419
+ course of action is usually immediately obvious.
420
+ On the other hand, if the situation does not seem
421
+ familiar complex RPD will be involved where the
422
+ decision maker adjusts the option after evaluating
423
+ it.Additional evidence was found in the specifictask domain of interest to the TADMUS program
424
+ which added support to these findings on the way
425
+ real-world decision makers make decisions in the
426
+ context of their normal jobs. Research was con-
427
+ ducted to determine decision requirements for
428
+ command-level decision makers in the combat
429
+ information center (CIC) of an Aegis cruiser.
430
+ Analysis of 14 incidents from actual problems re-
431
+ vealed 183 decisions. Of these, 103 concerned
432
+ situation assessments. Results obtained after ana-
433
+ lysts coded these situation assessments indicated
434
+ that decision makers arrived at approximately
435
+ 87% of their situation assessments through feature
436
+ matching and the remaining 13% through story
437
+ generation (Kaempf, Wolf, & Miller, 1993). The
438
+ other eighty decisions that were identified, from
439
+ analysis of the real-world incidents mentioned
440
+ above, involved course of action selection. These
441
+ course of action decisions served a variety of
442
+ functions, although, relatively few were intended
443
+ to end the incident. Twenty were intended as a
444
+ final course of action decision; 14 were imple-
445
+ mented to obtain more information, 22 to manage
446
+ resources, and 24 to put themselves in a more fa-
447
+ vorable tactical position. A recognition-based
448
+ strategy was also used by decision makers to de-
449
+ velop a course of action, accounting for 95% of
450
+ the actions taken in the 14 incidents. The decision
451
+ makers generated and compared multiple options
452
+ in only 5% of the cases. In line with these find-
453
+ ings, the TADMUS program has adopted the po-
454
+ sition that decision aiding systems should assist in
455
+ the decision making process, and focus on aidingthe situation assessment portion of the decision-making task.A DSS was developed to support decision-making processes which research has shown are
456
+ used by decision makers in real-world settings
457
+ (Hutchins, Kelly, & Morrison, 1996). Specifi-
458
+ cally, the DSS parallels the strategies used by ex-
459
+ perienced decision makers to perform situation
460
+ assessment (Nobel, 1989; 1993). This approach to
461
+ supporting the user's intuitive approach to dealing
462
+ with dynamic decision-making situations should
463
+ produce tools that are both more easily understood
464
+ and used, and that more effectively "exploit the
465
+ decision maker's knowledge and expertise that
466
+ might facilitate adaptation to complex, novel
467
+ situations" (Cohen, 1993, p. 265).4 Human-System Interaction PrinciplesThe vast majority of research on human-computerinteraction design has been devoted to character-
468
+ istics of displays that impact human perception,
469
+ such as symbol legibility or detectability, and on
470
+ relatively simple cognitive functions such as
471
+ memory tasks. Fewer efforts have been devoted to
472
+ understanding the effects of the format and man-
473
+ ner in which information is presented on more
474
+ complex levels of human cognition such as deci-
475
+ sion making. Consequently, principles that can be
476
+ applied to the design of the interface between the
477
+ user and a decision support system for the purpose
478
+ of enhancing cognitive changing situations, are
479
+ not available to any significant degree (ONT,
480
+ 1992).4.1 Graphic PresentationsSeveral advantages are offered by graphicpresentations over a text-based presentation for-
481
+ mat (Larkin and Simon, 1987). Graphic presenta-
482
+ tions should (1) reduce the amount of mental
483
+ computation required to perform tasks; and (2)
484
+ allow users to spend less time searching for
485
+ needed information. Casner (1991) elaborated on
486
+ these ideas and found that graphics allow users to
487
+ substitute less demanding perceptual opera-tions
488
+ for more complex logical operations. For exam-
489
+ ple, determining a change in altitude (and the de-
490
+ gree of change) is immediately apparent when the
491
+ user glances at the track history module. (Note
492
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 8Monterey, Ca.that the words contact and track can be used inter-changeably. The reader is referred to Figure 1.)The objective for the track history module isto facilitate the contact identification process by
493
+ providing information that is integrated in a way
494
+ that supports a recognitional decision strategy.
495
+ This module depicts a contact’s speed, altitude,
496
+ course and range on a two-dimensional graphical
497
+ display along with a geometric representation of
498
+ both the contact’s weapon release envelope and
499
+ own-ship’s weapons coverage. A large amount of
500
+ parametric data is portrayed graphically for rapidassimilation by the user. The user can see, at aglance, a synthesized picture of the contact’s be-
501
+ havior. Compare this rather simple perceptual op-
502
+ eration with the more complex logical operation
503
+ involved in current operational systems which re-
504
+ quire the user to recall and subtract numerical val-
505
+ ues for past and current altitudes.Graphics also allow users to omit steps thatare otherwise necessary when a task is performed
506
+ without a graphic. An example of this advantage
507
+ isFigure 1. Decision Support System Display Modules.also illustrated in the track history module whichincludes templates indicating weapon's coverage
508
+ for both the inbound contact and "own-ship." To
509
+ determine whether the aircraft is within its
510
+ weapon's launch range there is no need to recall
511
+ the specific launch range values and then compare
512
+ them with the aircraft's current range. Instead, the
513
+ user can determine if the aircraft is within its
514
+ launch range by a quick glance at the display.Graphics help users save time when searchingfor needed information when several related di-
515
+ mensions of information are encoded in a single
516
+ graphical object. This is accomplished by inte-
517
+ grating the kinematic parameters of speed, course,
518
+ altitude, bearing, and range for a contact. The user
519
+ can see, at a glance, a synthesized picture of the
520
+ contact's behavior. Compare this process with
521
+ reading, in a text-based format, the individual pa-
522
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 9Monterey, Ca.rameters which need to be integrated by the userinto a coherent picture of the contact's behavior.5 Limited Cognitive Processing CapabilitiesSince there are limits to the cognitive processingcapability of humans, it is important for the sys-
523
+ tem to provide the needed information in a format
524
+ that best supports the user operating under dy-
525
+ namic decision-making conditions. It may be the
526
+ case that current systems are inadequate to support
527
+ the cognitive processing demands required by
528
+ certain littoral scenarios. For example, according
529
+ to Gruner (1990, p. 41), the U.S.S. Vincennes of-
530
+ ficers and system operators "could not make better
531
+ decisions because they did not have time to con-
532
+ firm or deny the information uncertainties pre-
533
+ sented them." Gruner maintains that the rapid pace
534
+ involved in these types of situations can exceed
535
+ the capacity of the human to comprehend the
536
+ rapid flow of information presented by complex
537
+ systems. In the case of the Vincennes, the CIC
538
+ team had three minutes and 40 seconds to make
539
+ their decision. This includes the time required for
540
+ the operators to perceive and interpret sensor data
541
+ and for the commanding officer to make informed
542
+ judgments from these data (Roberts & Dotterway,
543
+ 1995). The result of the human's limited cognitive
544
+ processing capabilities is that the decision makers
545
+ may fail to remember critical pieces of data,
546
+ overlook stored information, draw hasty conclu-
547
+ sions, and produce flawed answers.Evidence of the effects of limitations in mem-ory and shared attention capacity on human deci-
548
+ sion making were found during baseline testing
549
+ (Hutchins & Kowalski, 1994; Hutchins & Westra,
550
+ 1995) and during empirical evaluation of the DSS
551
+ (Kelly, Hutchins, & Morrison, 1996; Kelly, Mor-
552
+ rison, & Hutchins. 1996).Simon (1978, p. 273) states, "...the human in-formation processing system...operates almost en-
553
+ tirely serially, one process at a time, rather than in
554
+ parallel fashion. This seriality is reflected in the
555
+ narrowness of its momentary focus of attention."
556
+ However, the AAW problem forces the decision
557
+ maker to operate in a parallel processing mode
558
+ when several contacts demand attention at the
559
+ same time. The requirement to monitor and
560
+ maintain an accurate SA for these concurrentcontacts, over the course of the evolving situation,imposes an additional load of strategically man-
561
+ aging the overall situation. Several researchers
562
+ have argued that "managing the attentional and
563
+ conceptual processes that permit cogent SA in-
564
+ volves significant cognitive resources" (Adams, et
565
+ al, 1995, p. 91; Endsley, 1988). The tasks of pri-
566
+ oritizing contacts and the associated actions to be
567
+ taken by the team, updating the status of critical
568
+ contacts, responding to the other requisite tasks in
569
+ the queue and, more generally, of strategically
570
+ managing the workload of current multitask sys-
571
+ tems under dynamically changing scenarios can
572
+ place an unrealistic cognitive load on the decision
573
+ maker.A major advantage offered by the experimen-tal DSS is that it should "buy time" for the user by
574
+ (1) performing many of the cognitive processing
575
+ tasks for the user and (2) by presenting informa-
576
+ tion in graphic format. The DSS will synthesize
577
+ much of the information used to develop situation
578
+ awareness and present a coherent picture of the
579
+ situation to the user. This integrated picture will
580
+ be portrayed graphically—rather than in the cur-
581
+ rent text-based format—which should further re-
582
+ duce the amount of time required to assimilate
583
+ this information. By performing several informa-
584
+ tion processing steps for the decision maker the
585
+ decision maker's limited cognitive resources can
586
+ be used for the types of decisions which require
587
+ human abilities (e.g., the decision on whether to
588
+ engage).5.1 Working Memory RequirementsAn essential information processing step re-quired by this task—and one which levies a heavy
589
+ load on working memory—involves integrating
590
+ kinematic and sensor variables and maintaining an
591
+ awareness of changes in these variables over time.
592
+ Changes in a contact's behavior such as, decreas-
593
+ ing altitude, increasing speed, changes in elec-
594
+ tronic emissions, etc., can provide key indicators
595
+ of possible hostile intent. With current systems,
596
+ the decision maker receives numerous reports
597
+ from CIC team members who provide various
598
+ pieces of the overall tactical picture (such as,
599
+ kinematic parameter values, active electronic
600
+ emitter identifications, and behavioral responses
601
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 10Monterey, Ca.of the contact in response to queries by the team)regarding a particular contact. Some of this infor-
602
+ mation is also displayed in a text-based format for
603
+ the user when a contact is "hooked" (that is, se-
604
+ lected for display) by the decision maker. How-
605
+ ever, to recognize a change in certain variables,
606
+ current systems require the user to retain parame-
607
+ ter values in short-term memory in order to recog-
608
+ nize a change in the parameter, such as altitude.When the decision maker is monitoring sev-eral concurrent contacts (such as, cycling through
609
+ three or four contacts in a 1-minute period) human
610
+ working memory capabilities may quickly be sur-
611
+ passed. To detect a change in a critical parameter
612
+ value, the decision maker must maintain the pa-
613
+ rameter values for the contacts of interest in
614
+ working memory as he or she cycles between sev-
615
+ eral contacts. For example, the decision maker
616
+ must be able to recall that contact 7022 was at
617
+ 14,000 ft. altitude one minute ago, and then sub-
618
+ tract the current altitude value of 10,000 ft., which
619
+ will then indicate the aircraft is in a rapid descent.
620
+ The DSS was developed to aid the decision maker
621
+ by performing several of these cognitive process-
622
+ ing tasks, thus, reducing the cognitive load for the
623
+ user. By presenting the synthesized picture of the
624
+ contact's behavior over time, through the use of
625
+ graphical displays, critical changes should be
626
+ immediately apparent to the user.A second memory-intensive task involvesmaintaining, in working memory, a current list of
627
+ actions taken by team members, the contact's re-
628
+ sponse to these actions taken by the CIC team,
629
+ and pending actions. Research has established that
630
+ "memory is limited and that list maintenance is
631
+ effortful and fallible—more so if the list must be
632
+ ordered and still more if the membership of the
633
+ list must be dynamically reordered and modified
634
+ during retention" (Bower, 1970, as cited in Ad-
635
+ ams, et al, 1995, p. 91). The DSS should reduce
636
+ the cognitive effort required for distributing atten-
637
+ tion among the many contacts to be attended to
638
+ and actions that are required. Working memory
639
+ requirements should be reduced by having the
640
+ DSS act as an intelligent "assistant," reminding
641
+ the user regarding what actions are to be taken and
642
+ when the actions are to be taken.A third way the DSS will reduce memory andinformation processing requirements is by dis-
643
+ playing templates depicting weapons' envelopes
644
+ for both the inbound contact and "own-ship." This
645
+ should facilitate critical comparisons and judg-
646
+ ments regarding timing of actions. During a sce-
647
+ nario decision makers have to either rely on mem-
648
+ ory to recall the launch range for various weapons
649
+ or query a team member for this information. Both
650
+ of these methods waste limited resources. The
651
+ high workload and high tempo characteristic of
652
+ littoral scenarios produce a stress-ful decision-
653
+ making environment. The phenom-enon that in-
654
+ creasing stress leads to decreasing working mem-
655
+ ory is well documented (e.g., Hockey, 1986). The
656
+ latter method for obtaining the desired informa-
657
+ tion wastes limited resources by increasing the
658
+ communications load and requiring more time to
659
+ wait for a team members' response to the queryUnder these high-tempo and high workloadconditions human memory and attentional re-
660
+ sources can easily be surpassed. Several cogni-
661
+ tively resource intensive information processing
662
+ steps are eliminated for the human decision maker
663
+ by having them performed by the DSS. We pre-
664
+ dict that the decision support tools will reduce the
665
+ cognitive workload imposed on the decision
666
+ maker in the following three ways: (1) by reduc-
667
+ ing the amount of information processing to be
668
+ performed, (2) reducing working memory re-
669
+ quirements, and (3) assisting the user in allocating
670
+ limited attentional resources.5.2 Reducing Human ErrorThe study of human cognitive processes andrelated error mechanisms has gained rapidly in-
671
+ creasing interest in the past decade. Rasmussen
672
+ (1987) argues that the emphasis in attempting to
673
+ understand human errors must shift from tasks to
674
+ the human-task mismatch. For example, Gruner
675
+ (p. 39), in discussing the Vincennes incident,
676
+ maintains that "the system was poorly suited for
677
+ use by human beings during rapid military action."
678
+ He ascribes this lack of suitability to a human-
679
+ machine mismatch between the rate of data flow
680
+ possible with modern computer systems that can
681
+ process and display information at phenomenal
682
+ data rates and the "comprehension capability of
683
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 11Monterey, Ca.users which has remained almost static for thou-sands of years." This causal approach to under-
684
+ standing human error is based on the premise that
685
+ errors are rarely random and can be traced to
686
+ causes and contributing factors. Once these con-
687
+ tributing factors are identified they can be miti-
688
+ gated.The impact and vulnerability of systems andhuman interfaces, because of incompatibilities
689
+ between the way people perceive, think, and act,
690
+ are documented in the popular and technical lit-
691
+ erature (Buck, 1989; Casey, 1993; Norman, 1988;
692
+ Perrow, 1984; Wilson & Zarakas, 1978). Newly
693
+ developed systems will succeed or fail based on
694
+ our ability to minimize these incompat-ibilities
695
+ between the characteristics of the things we create
696
+ and the way we use them. There are many well-
697
+ documented instances of critical systems or pa-
698
+ rameter changes going unnoticed or unheeded be-
699
+ cause the operating procedures, or the human ma-
700
+ chine interface, provided no historical trace. For
701
+ example, an unnoticed increase in altitude con-
702
+ tributed to the shoot down of the Iranian airbus by
703
+ a U.S. Navy ship—when the team mistakenly be-
704
+ lieved the aircraft to be descending—because
705
+ there was no historical trace to make the aircraft's
706
+ actual increasing altitude apparent (Dotterway,1992). Five personnel in the U.S.S. Vincennes's
707
+ combat information center, all viewing separate
708
+ displays, reported the aircraft as descending while
709
+ the Aegis data tapes later revealed a flight pattern
710
+ of ascent (Roberts & Dotterway, 1995). One of the
711
+ official investiga-tions of this incident, the Fo-
712
+ garty Report (1988, p. 45), states that "stress, task
713
+ fixation, and an unconscious distortion of data
714
+ may have played a major role in this incident." A
715
+ panel of five psychologists from the American
716
+ Psychological Association who testified before
717
+ Congress concluded that there were "predictable
718
+ failings of human judgment under intense stress
719
+ compounded by complex technology [which]
720
+ clearly contributed to the accidental shooting of
721
+ Iranian airliner Flight 655" (APA, p. 4).It is generally accepted that between 60-80percent of the accidents and malfunctions in
722
+ transportation, manufacturing, process control,
723
+ weapon, and other systems are attributable to hu-
724
+ man error (Senders & Moray, 1991; Van Cott,1993; Weiner, 1994). Reducing tactical decisionmaking errors is one goal of the TADMUS pro-
725
+ gram. The following section presents a brief re-
726
+ view of an experiment conducted to develop a
727
+ baseline on tactical decision making performance
728
+ in response to fairly stressful scenarios. A com-
729
+ panion paper (Hutchins, Kelly, & Morrison, 1996)
730
+ describes the experimental DSS modules and the
731
+ way they are hypothesized to enhance tactical de-
732
+ cision making performance.6 TADMUS Baseline ExperimentEarly research involved data collection in the De-cision-Making Evaluation Facility for Tactical
733
+ Teams (DEFTT) Laboratory using simulated ex-
734
+ isting shipboard displays to establish a baseline on
735
+ decision-making performance. The purpose of this
736
+ effort was to document baseline decision-making
737
+ performance for experienced naval officers. Dur-
738
+ ing the baseline phase of testing, a detailed under-
739
+ standing was developed of the cognitive processes
740
+ underlying the various tasks involved in situation
741
+ assessment—and where the bottlenecks occur.
742
+ This understanding was then used to design the
743
+ way the information is presented to the user in or-
744
+ der to facilitate performance of the required tasks.6.1 SubjectsThis study focused on the command-level de-cision makers of an antiair warfare team on an
745
+ Aegis cruiser—the commanding officer and the
746
+ tactical action officer. Subjects in the study con-
747
+ sisted of six commanding officer/tactical action
748
+ officer teams drawn from twelve active duty Na-
749
+ val personnel; some were from training com-
750
+ mands while others were from operational com-
751
+ mands aboard ship or assigned to group staffs.6.2 ProcedureData were collected in the DEFTT Labora-tory, a six-station test-bed environment that
752
+ simulates console positions in a Navy Aegis
753
+ cruiser combat information center. (For a detailed
754
+ description of the DEFTT Laboratory see Hutch-
755
+ ins, 1996.) Four stations were filled by confeder-
756
+ ates (active duty Navy personnel) who play antiair
757
+ warfare support-team member roles. These roles
758
+ included the antiair warfare coordinator, identifi-
759
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 12Monterey, Ca.cation supervisor, tactical information coordina-tor, and electronic warfare supervisor. After ap-
760
+ proximately 1 1/2 hours of orientation to the labo-
761
+ ratory and training in the use of the computer con-
762
+ soles the subjects engaged in four scenarios. The
763
+ scenarios were each about 25 minutes in length
764
+ and contained between 11 and 14 contacts of in-
765
+ terest per scenario, in addition to numerous back-
766
+ ground contacts.6.3 Treatment of DataTeam communications were recorded on amultichannel audio recorder; these included all
767
+ intra-team exchanges, as well as all communica-
768
+ tions with simulated off-ship personnel. Audio
769
+ tapes were used to produce verbatim, time-
770
+ stamped transcripts of all team communications.
771
+ A modified version of the TapRoot® Incident In-
772
+ vestigation System (Paradies, 1991; Paradies and
773
+ Unger, 1991) was then applied to identify errors.
774
+ The objective was to identify tactically significant
775
+ errors committed during the scenario. Tactically
776
+ significant errors were defined as those errors that
777
+ may lead to loss of life or significant political em-
778
+ barrassment. The following criteria were used for
779
+ counting an error as tactically significant: (1) loss
780
+ of situation awareness, (2) failure to take defen-
781
+ sive action when within the weapon's range of an
782
+ approaching contact, or (3) a violation of rules of
783
+ engagement (ROE). Video recordings were made
784
+ of the commanding officer and tactical action of-
785
+ ficer computer screens. Detailed analyses of all
786
+ audio and video recordings were conducted. (For
787
+ a more detailed coverage of the methodology and
788
+ results see Hutchins and Westra, in preparation).6.4 ResultsThe complex, time-constrained, decision-making situations embodied in the experimental
789
+ scenarios resulted in a large number of decision
790
+ errors. The mean number of tactically significant
791
+ errors documented across six teams and four sce-
792
+ narios was 14; the number of errors ranged from
793
+ nine to twenty-two. The standard deviation was
794
+ 3.7. Subjects performed an average of 50% of the
795
+ required behaviors as specified in the rules of en-
796
+ gagement. The ordinal agreement between three
797
+ raters (navy subject matter experts) on error countranks from TapRoot® analyses was computed.Results showed a high degree of agreement with
798
+ the Kendall's W of .93 indicating that 93% of the
799
+ possible rank variance is accounted for.6.4.1 Decision-Making ErrorsDetailed examinations of the informationprocessing sequences performed during tactical
800
+ decision making have revealed a variety of errors.
801
+ On average, subjects failed to take required ac-
802
+ tions, about half of the time. Explanations based
803
+ in the cognitive psychology literature have been
804
+ pursued, as a major goal of the TADMUS pro-
805
+ gram is to develop a DSS based on an under-
806
+ standing of the way in which human decision
807
+ makers actually process information under rapidly
808
+ evolving situations.The majority of documented errors involvederrors of omission, that is, "failure to take defen-
809
+ sive measures” and "failure to adhere to ROE.”
810
+ Failure to take defensive measures included fail-
811
+ ure to take actions to defend own-ship when an
812
+ approaching aircraft had reached its weapon's re-
813
+ lease range. An example involved a case where
814
+ two contacts were within the specified ROE limit,
815
+ yet no action had been taken. The types of actions
816
+ included in the “failure to adhere to ROE” cate-
817
+ gory include failure to take action regarding the
818
+ items listed and defined below: (a) issuing warn-ings is part of the usual identification process andinvolves three levels of warnings with increasing
819
+ levels of urgency; (b) establish friendly force cri-teria refers to establishing a plan with otherfriendly ships in the area to coordinate how they
820
+ will respond to potential threats; (c) changes inkinematics/ identification friend or foe—subjectsare expected to notice significant kinematic
821
+ changes and/or identification friend or foe pa-
822
+ rameter changes; and (d) other identification pro-cedures includes actions such as illuminating withfire control radar.The other major category of tactically signifi-cant error involved “loss of SA.” Loss of SA er-
823
+ rors were grouped under errors of commission and
824
+ errors of omission and then further categorized
825
+ into subgroups. Fifty-five percent of the loss of
826
+ SA errors involved taking the wrong action (error
827
+ of commission) while 45% of the errors involved
828
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 13Monterey, Ca.failing to take some required action (error ofomission). Error categories included under errors
829
+ of commission involved incorrectly engaging a
830
+ track (3%), incorrectly warning a track (29%),
831
+ other incorrect actions (16%), and incorrect re-
832
+ porting (7%). The two instances of incorrectly en-
833
+ gaging an aircraft, which were F-1 Mirage aircraft,
834
+ were considered errors because the decision
835
+ maker failed to take certain actions prior to en-
836
+ gaging—not necessarily because the aircraft
837
+ should not have been engaged. The actions that
838
+ the decision makers failed to take involved ascer-
839
+ tain-ing the identification of the aircraft for one
840
+ case and failure to warn and illuminate the aircraft
841
+ prior to engaging for the second case. Most in-
842
+ stances of incorrectly issuing warnings to the air-
843
+ craft involved issuing the warning when the air-
844
+ craft was within its territorial airspace (that is, in-
845
+ side the 12 nautical mile limit which is interna-
846
+ tionally recognized as under control of that nation)
847
+ or issuing a warning at a level different from what
848
+ was required. Other incorrect actions included il-
849
+ luminating the aircraft, “locking up” with radar, or
850
+ ordering the aircraft to divert when the aircraft
851
+ was still within its territorial airspace. Incorrect
852
+ reporting involved inaccurate reports on the status
853
+ of the tactical situation (such as, indicating to the
854
+ battle group commander that certain actions had
855
+ been taken when they had not, misidentification of
856
+ an aircraft, or omitting critical tracks from a re-
857
+ port).Errors of omission categorized under the “lossof SA” category included: (a) failure to identify or
858
+ attend to a contact; (b) failure to take action (e.g.,
859
+ to issue “hold-fire” when a contact turned out-
860
+ bound); (c) failure to recognize a threat (e.g.,
861
+ designating an aircraft as a non-threat because it
862
+ had passed its closet point-of-approach, yet it was
863
+ still within missile-launch range); (d) instances of
864
+ confusion or forgetting (e.g., forgetting or ignor-
865
+ ing critical data, forgetting whether or not it had
866
+ been warned, illuminated, or “locked-on,” or for-
867
+ getting the aircraft's response, or lack of response
868
+ to these actions, forgetting the status of a contact,
869
+ and confusing contacts); (e) misperception of data
870
+ (e.g., reporting a contact as turning outbound
871
+ when it is still inbound); (f) unclear communica-
872
+ tion (issuing vague orders regarding actions to betaken by team member, such as, failure to specifywhich weapon system is to be used or which con-
873
+ tact is to be engaged).6.4.2 Cognitive explanationsThe cause of failures to take required actionsis, in many cases, attributed to the extremely high
874
+ task demands levied on the decision maker by the
875
+ scenario and the human decision-maker's limited
876
+ attentional resources. Many cases are also attrib-
877
+ uted to working memory limitations. Maintaining
878
+ an awareness of the status of each contact and the
879
+ status of many actions to be taken by the antiair
880
+ warfare team—which actions have been taken and
881
+ what the contact's response to the action was—
882
+ severely taxes the decision maker's working
883
+ memory. The high workload entailed in the
884
+ scenarios produces a highly time-compressed
885
+ decision making situation. This time-compressed
886
+ decision making situation—where attentional
887
+ resources and working memory capacity are
888
+ limited—do not allow the decision maker to
889
+ maintain accurate SA for all tracks at any given
890
+ time. We anticipate that the decision support
891
+ modules in the DSS will mitigate these types of
892
+ errors.Human information processing capabilities arenot well suited to dealing with a "multiplicity of
893
+ simultaneous and disjointed tasks. Thoughtful at-
894
+ tention is modular: People can consciously think
895
+ about only one thing at a time" (Adams, et al,
896
+ 1995, p. 92). As a result, they do not handle inter-
897
+ ruptions very well. Research indicates that when
898
+ an operator is faced with as few as two tasks that
899
+ consist of merely the detection or recognition of
900
+ simple signals, a cost may be incurred in terms of
901
+ a significant loss in sensitivity or time that can be
902
+ allocated to either by the requirement to divide or
903
+ switch attention between them (Broadbent, 1957;
904
+ Schneider and Detweiler, 1988; Swets, 1984).The memory demands of managing complex,multi-task situations can easily surpass human
905
+ limitations. The decision maker must not forget
906
+ any of the contacts or tasks requiring action. In
907
+ addition to remembering all the tasks needing at-
908
+ tention, however, are the complexities entailed in
909
+ keeping track of the data and substeps associated
910
+ with each contact and prior action. The aviation
911
+ literature provides many examples of incidents
912
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 14Monterey, Ca.with explanations similar to the root causes forerrors that were found in the TADMUS program.
913
+ One category includes the potentially disastrous
914
+ effects of interruptions in the task for air traffic
915
+ controllers and pilots. Similarly, in the AAW en-
916
+ vironment, momentary intervening attention to
917
+ another task or contact, or an interruption in a
918
+ procedure can leave the procedure, or processing
919
+ of a contact incomplete with potentially cata-
920
+ strophic results.A fairly consistent pattern of tactical decision-making errors was documented from data col-
921
+ lected during the baseline data collection period.
922
+ The root causes of these errors were traced to
923
+ cognitive mechanisms such as limited attentional
924
+ resources and working memory limitations. By
925
+ developing an understanding of the pattern and
926
+ types of errors most frequently observed in this
927
+ task domain we hope to provide a DSS which will
928
+ mitigate these errors.7 DiscussionFailure to take appropriate actions may be ex-plained by the limited resource capacity of human
929
+ memory. In these scenarios a large number of
930
+ contacts are monitored for changes in any of sev-
931
+ eral key parameters. Three modules in the DSS
932
+ are hypothesized to assist with recognizing a
933
+ problem and taking the appropriate actions: track
934
+ history; response manager; and the track priority
935
+ list and alerts.Features offered by the DSS to address errorsattributed to limited attentional resources include
936
+ focusing attention on (1) high priority contacts
937
+ (i.e., track priority list and alerts), as well as on (2)
938
+ missing data (e.g., basis for assessment), and (3)
939
+ enabling the decision maker to use more data than
940
+ is typically used in current systems (e.g., track
941
+ history, comparison to norms). Current systems
942
+ require the user to retain previous contact data in
943
+ memory to compare with current values for criti-
944
+ cal parameters. Current systems also require the
945
+ user to rely on recall of vast amounts of informa-
946
+ tion from training and experience. Presenting all
947
+ known data on a contact in a synthesized way
948
+ should reduce working-memory requirements and
949
+ facilitate recognition. Additional potential per-
950
+ formance enhancement features, offered by theDSS, include displaying the complete kinematiccontact history, presenting graphic displays of lo-
951
+ cation and trends, highlighting missing data, pro-
952
+ viding alerts, and providing assessments of cur-
953
+ rent contact identity that go beyond what existing
954
+ systems currently present.Focusing the user's attention on trend andhistory data should decrease the cognitive work-
955
+ load imposed by these scenarios where many
956
+ contacts must be identified and responded to un-
957
+ der severe time constraints. Similarly, delineating
958
+ trend and history data can assist in the identifica-
959
+ tion of a contact where noticing changes in critical
960
+ parameters is essential. Presentation of trend and
961
+ history data, as well as threat assessment and
962
+ comparison to norms, should also mitigate cogni-
963
+ tive "tunnel vision" effects where the decision
964
+ maker attends to a smaller number of cues when
965
+ under stress.The notion of time is an important character-istic of situation awareness (Harwood, Barnett,
966
+ and Wickens, 1988). The past is critical to under-
967
+ standing the present, and both past and present
968
+ information must be used to predict future events
969
+ (Shrestha, et al, 1995). Endsley (1988) referred to
970
+ the "projection of their (perceived elements) status
971
+ in the near future" when discussing situational
972
+ awareness. However, Endsley also noted the task
973
+ of attending to incoming information and subse-
974
+ quently predicting future events places a heavy
975
+ load on working memory. Several decision sup-
976
+ port modules were developed to assist the user in
977
+ remaining aware of the contact's history and
978
+ changes over time. Remembering which actions
979
+ are to be taken at what time levies an additional
980
+ burden by placing a heavy load on working mem-
981
+ ory. A secondary time savings should be achieved
982
+ by the DSS acting as an intelligent "advisor," that
983
+ is, by assisting the decision maker in knowing
984
+ what actions to take, when to take them, and
985
+ which actions have already been taken. A tertiary
986
+ time savings can be achieved by including a tem-
987
+ plate depicting the weapons' release ranges so the
988
+ decision maker does not need to rely on fallible
989
+ human memory or query a team member regard-
990
+ ing weapons ranges. By graphically depicting a
991
+ synthesized view of a contact's kinematic history,
992
+ with the focus on changes in the contact's behav-
993
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 15Monterey, Ca.ior over time, along with the contact's weaponsenvelope in relation to both own-ship's radar and
994
+ weapon coverage, information processing time
995
+ can be saved for the decision maker.8 ConclusionsThe research reported here focuses on developinga DSS which reflects the natural decision-making
996
+ strategies of humans. Presenting synthesized in-
997
+ formation in the form of graphic presentations is
998
+ expected to reduce the cognitive processing load
999
+ for the decision maker when performing situation
1000
+ assessment. The intention is to aid the decision
1001
+ maker by providing information in a way that will
1002
+ minimize the need to maintain information in
1003
+ working memory, reduce information processing
1004
+ demands, help focus attentional resources on the
1005
+ highest priority contacts, remind the user of ac-
1006
+ tions which need to be taken, help make decisions
1007
+ under stress, and support higher levels of situation
1008
+ awareness.Decision support systems require that the hu-man's strengths be used in synergy with the ad-
1009
+ vantages offered by the DSS. Limitations associ-
1010
+ ated with the current generation of automated de-
1011
+ cision aids include the idea that (1) they cannot
1012
+ adequately capture the expertise developed by ex-
1013
+ perience over time and (2) since all contingencies
1014
+ cannot be anticipated, the expert's abilities to use
1015
+ intuition is indispensable (Mosier, in press). Mo-
1016
+ sier's review of the limitations of automated deci-
1017
+ sion systems delineates the characteristics of hu-
1018
+ man expertise that surpass the capabilities of
1019
+ automated systems. These include the human ca-
1020
+ pacity for creativity, adaptability, the ability to
1021
+ incorporate experience, the presence of a broad
1022
+ focus, analogical reasoning, and commonsense
1023
+ knowledge. The goal for the DSS is to capitalize
1024
+ on the strengths of the human along with the ad-
1025
+ vantages provided by the decision support system.9 Testing the DSSThe prototype DSS display modules are currentlybeing empirically evaluated in the simulated tacti-
1026
+ cal environment provided in the DEFTT Labora-
1027
+ tory. Experienced naval decision makers engage
1028
+ in experimental scenarios with and without access
1029
+ to the DSS display. The various decision supportmodules will be tested individually and in combi-nation in future experiments. Data on reduction of
1030
+ errors, improvements in users' situation awareness
1031
+ scores, changes in communication patterns, and
1032
+ subjective responses to the decision support sys-
1033
+ tem will be collected.10 Future ResearchWhile tools based on both the RPD and explana-tion-based reasoning models of decision making
1034
+ are included in the DSS there is no direct connec-
1035
+ tion between the two. Research is currently being
1036
+ conducted to extend schema theory to dynamic
1037
+ decision-making situations. This involves devel-
1038
+ oping and testing a hybrid model of cognitive be-
1039
+ havior in decision making to incorporate both
1040
+ types of knowledge, i.e., feature matching and
1041
+ story generation, as elements of the same schema
1042
+ model of naturalistic decision making (Smith &
1043
+ Marshall, in press). Schema theory as described
1044
+ by these authors, offers a context for integrating
1045
+ these two models which have typically been
1046
+ viewed as separate entities.AcknowledgmentsThe authors gratefully acknowledge the assistanceof Steve Francis, Brent Hardy, C.C. Johnson, Pat
1047
+ Kelly, Ron Moore, Connie O’Leary, Pat Marvel,
1048
+ Mike Quinn, and Will Rogers in data collection,
1049
+ interpretation, developing the DSS, and in con-
1050
+ ducting this research.ReferencesAdams, M. J., Tenney, Y. J., & Pew, R. W.(1995). Situation Awareness and the Cognitive
1051
+ Management of Complex Systems. HumanFactors, 37(1), 85-104.American Psychological Association, ScienceAgenda, Fall, 1988.Bower, G. H. (1970). Organizational Factors inMemory. Cognitive Psychology, 1, 18-46.Broadbent, D. E. (1957). A mechanical model forhuman attention and immediate memory. Psy-chological Review, 64, 205-215.Buck, L. (1989). Human error at sea. HumanFactors Bulletin, September, 32 (9), 12.
1052
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 16Monterey, Ca.Casey, S. (1993). Set Phasers on Stun and OtherTrue Tales of Design, Technology, and Hu-
1053
+ man Error. Aegean Publishing Company,Santa Barbara, CA.Casner, S. (1991). Task-Analytic Design ofGraphic Presentations. Technical Report AIP-145. Departments of Computer Sci-ence and Psychology, Carnegie Mellon Uni-
1054
+ versity and Learning Research and Devel-
1055
+ opment Center, University of Pittsburgh, PA.Cohen, M. S. (1993). The Bottom Line: Natu-ralistic Decision Aiding. In G. Klein, J. Ora-
1056
+ sanu, R. Calderwood, & C. E. Zsambok (Eds.)Decision Making in Action: Models andMethods (pp. 138-147). Ablex PublishingCorporation, New Jersey.Dawes, R. M. (1971). A case study of graduateadmissions: Application of three principles of
1057
+ human decision making. American Psycholo-gist. 34, 571-582.Dotterway, K. A. (1992). Systematic Analysis ofComplex Dynamic Systems: The case of the
1058
+ USS Vincennes. Unpublished master's thesis,Naval Postgraduate School, Monterey, CA.Endsley, M. R. (1988). Design and evaluation forsituational awareness enhancement. Pro-ceedings of the Human Factors 32nd Annual
1059
+ Meeting, 97-101.Federico, P. A. (1995). Expert and novice recog-nition of similar situations. Human Factors,37(1), 105-122.Formal Investigation into the Circumstances Sur-rounding the Downing of a Commercial Air-
1060
+ line by the USS Vincennes, by Rear Admiral
1061
+ William M. Fogerty, USN, 28 July 1988, pp.
1062
+ 4, 6. Hereafter referred to as the Fogerty Re-
1063
+ port.Gruner, W. P. No Time For Decision Making.(1990, November). U.S. Naval Institute Pro-ceedings, 39-41.Harwood, K., Barnett, B., and Wickens, C.(1988). Situational awareness: A conceptual
1064
+ and methodological framework. Proceedingsof the Symposium Psychology in the Depart-
1065
+ ment of Defense.Hockey, G. R. (1986). Changes in Operator Effi-ciency as a Function of Environmental Stress,
1066
+ Fatigue, and Circadian Rhythms. In: K. R.Boff, L. Kaufman, J. P. & Thomas (Eds.):Handbook of Perception and Human Per-formance. Wiley, New York.Hutchins, S. G. (in press). Decision-MakingEvaluation Facility for Tactical Teams. Naval
1067
+ Command, Control, and Ocean Surveillance
1068
+ Center, RDT&E Division Technical Report, in
1069
+ press, San Diego, CA.Hutchins, S. G., Kelly, R. T. and Morrison, J. G.(in press). Decision Support for Tactical Deci-
1070
+ sion Making Under Stress. Proceedings of theSecond International Symposium on Com-
1071
+ mand and Control Research and Technology.June 1996, Monterey, CA.Hutchins, S. G. and Kowalski, J. T. (1993). Tac-tical Decision Making Under Stress: Prelimi-
1072
+ nary Results and Lessons Learned. Proceed-ings of the 10th Annual Conference on Com-
1073
+ mand and Control Decision Aids. June 1993,Washington, D. C.Hutchins, S. G. and Rummel, B. K. (1995). ADecision Support System for Tactical Deci-
1074
+ sion Making Under Stress. Proceedings of the
1075
+ First International Conference on Command
1076
+ and Control Research and Technology. June
1077
+ 1995, Washington, D. C.Hutchins, S. G. and Westra, D. P. (1995). Patternsof Errors Shown by Experienced Navy Com-
1078
+ bat Information Center Teams. Proceedingsof the 39th Annual Meeting of the Human
1079
+ Factors and Ergonomics Society, San Diego,CA. October 1995.Hutchins, S. G. and Westra, D. P. (in preparation).TADMUS Baseline Experimental Results.
1080
+ Naval Command, Control, and Ocean Sur-
1081
+ veillance Center, RDT&E Division Technical
1082
+ Report, in preparation, San Diego, CA.Kaempf, G. L. and Militelo, L. G., (1992). Deci-sion Making in Emergencies, First Offshore
1083
+ Installation Management Conference: Emer-
1084
+ gency Command Responsibilities Collected
1085
+ Papers, Aberdeen, Scotland.Kaempf, G. L. , Wolf, S. and Miller, T. E. (1993).Decision Making in the Aegis Combat Infor-
1086
+ mation Center. In Proceedings of the Human
1087
+ Factors and Ergonomics Society 37th Annual
1088
+ Meeting (pp. 1107-1111). Santa Monica, CA:
1089
+ Human Factors and Ergonomics Society.
1090
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 17Monterey, Ca.Kelly, R. T. , Hutchins, S. G., and Morrison, J. G.(1996). Decision Processes and Team Com-
1091
+ munications with a Decision Support System.Proceedings of the Second International Sym-posium on Command and Control Research
1092
+ and Technology. June 1996, Monterey, CA.Kelly, R. T. , Morrison, J. G., and Hutchins, S.G., (in press). Impact of Naturalistic Decision
1093
+ Support on Tactical Situation Awareness.Proceedings of the 40th Annual Meeting of theHuman Factors and Ergonomics Society,Philadelphia, PA. September 1996.Kirshenbaum, S. S. (1992). Influence of Experi-ence on Information-Gathering Strategies.Journal of Applied Psychology, 77, 343-352.Klein, G. A. (1989). Recognition-Primed Decisions. In W. R. Rouse (Ed.) Advancesin Man-Machine Systems Research (pp. 47 - 92), Vol. 5. JAI Press, Inc.Klein, G. A. (1993). A Recognition-Primed Deci-sion (RPD) Model of Rapid Decision Making.
1094
+ In G. A. Klein, J. Orasanu, R. Calderwood, &
1095
+ C. E. Zsambok (Eds.) Decision Making inAction: Models and Methods (pp. 138-147).Ablex Publishing Corporation, New Jersey.Larkin, J. H. and Simon, H. A. (1987). Why adiagram is (sometimes) worth 10,000 words.Cognitive Science, 1, 65-99.Mosier, (in press). Myths associated with Auto-mated Decision Aids. In G. A. Klein, & C. E.
1096
+ Zsambok (Eds.) Advances in Naturalistic De-cision Making: Research and Applications,Hillsdale, NJ: Erlbaum.Mundy, C. E., Jr. (1994). Thunder and Light-ning: Joint Littoral Warfare, Joint ForceQuarterly, 4, Spring, 45-50.Noble, D. (1989). Application of theory of cog-nition to situation assessment. Vienna, VA:Engineering Research Associates.Noble, D. (1993). A Model to Support Devel-opment of Situation Assessment Aids. n G.
1097
+ A. Klein, J. Orasanu, R. Calderwood, & C. E.
1098
+ Zsambok (Eds.) Decision Making in Action:Models and Methods (pp. 287-305). AblexPublishing Corporation, New Jersey.Norman, D. A. (1988). The Psychology of Eve-ryday Things. Basic Books, Inc. New York.Norman, D. A. and Bobrow, D. G. (1975). Onthe data-limited and resources-limited proc-
1099
+ esses. Cognitive Psychology, 7, 44-64.Office of Naval Technology. (1992). FY 1993Program Plan for Tactical Decision-Making
1100
+ Under Stress, Arlington, VA: July 1992.Paradies, M. (1991). Root Cause Analysis andHuman Factors. Human Factors Society Bul-letin, 34(8), 1-4.Paradies, M. & Unger, L. (1991). TapRoot Inci-dent Investigation System Manual. Volumes1-7. System Improvements, Inc. Knoxville,
1101
+ TN.Pennington, N. & Hastie, R (1992). Explainingthe Evidence: Tests of the Story Model of De-
1102
+ cision Making. Journal of Personality and So-cial Psychology, Vol. No. 2, 189-206.Pennington, N. & Hastie, R (1993). A theory ofExplanation-Based Decision Making. In G. A.
1103
+ Klein, J. Orasanu, R. Calderwood, & C. E.
1104
+ Zsambok (Eds.) Decision Making in Action:Models and Methods (pp. 188-201). AblexPublishing Corporation, New Jersey.Perrow, C. (1984). Normal Accidents: Livingwith High Risk Technologies. New York, Ba-sic Books, Inc.Rasmussen, J. (1986). Information Processingand Human-Machine Interaction. In A. P.
1105
+ Sage (Ed.) Series Volume 12, North-Holland,
1106
+ Amsterdam.Roberts, N. C. & Dotterway, K. A. (1995). TheVincennes Incident: Another Player on theStage? Defense Analysis Vol 11, No. 1, pp.31-45.Salthouse, T. A. (1992). Cognition and Context.Science, 257, 982-983.Sarter, N. B. and Woods, D. D. (1991). Situationawareness: A critical but ill-defined phe-
1107
+ nomenon. The International Journal of Avia-tion Psychology, 1(1), 45-57.Schneider, W., and Detweiler, M. (1988). Therole of practice in dual-task performance:
1108
+ Toward workload modeling in a connectionist/
1109
+ control architecture. Human Factors, 30, 539-566.Senders, J. W. & Moray, N. P. (1991). HumanError: Cause, Prediction, and Reduction.Lawrence Erlbaum Associates, New Jersey.
1110
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 18Monterey, Ca.Shrestha, L. B., Prince, C., Baker, D. P. and Salas,E., (1995). Understanding Situation Aware-
1111
+ ness: Concepts, Methods, and Training. Hu-man/Technology Interaction in Complex Sys-
1112
+ tems. Vol 7, W. B. Rouse (Ed.). San Fran-cisco: JAI Press.Simon, H. A. (1978). Information-processingtheory of human problem solving. In W. K.
1113
+ Estes (Ed.), Handbook of Learning and Cog-nitive Processes, Vol 5, Human InformationProcessing, New York: Wiley.Smith, D. E. and Grossman, J. D. (1993). Under-standing and Aiding Decision Making in
1114
+ Time-Constrained and Ambiguous Situations.Unpublished manuscript.Smith, D. E. and Marshall, S. (in press). Apply-ing Hybrid Models of Cognition in Decision
1115
+ Aids. In G. A. Klein, & C. E. Zsambok (Eds.)Advances in Naturalistic Decision Making:Research and Applications, Hillsdale, NJ:Erlbaum.Swets, J. A. (1984). Mathematical models of at-tention. In R. Parasuraman and R. Davies
1116
+ (Eds.), Varieties of Attention (pp. 183-242).New York: Academic.Tolcott, M. A. (1991). Understanding and AidingMilitary Decisions. Office of Naval ResearchEuropean Office. 27th International Applied
1117
+ Psychology Symposium, Stockholm, Sweden,
1118
+ June 1991.Wilson, G. L. & Zarakas, P. (1978). Anatomy ofa blackout. IEEE Spectrum, February, 339-346.
1119
+ Paper presented to the Second International Command and Control Research and Technology Symposium, 19Monterey, Ca.The requirement to interleave a mul-tiplicity of tasks—although not necessary
1120
+ an ongoing characteristic of shipboard
1121
+ scenarios—represents the type of situation
1122
+ where providing decision support may
1123
+ make the critical difference in the outcome
1124
+ for a scenario. For example, during the
1125
+ experimental scenarios the decision mak-
1126
+ ers may have to perform the following: monitor ship location (relative toother ships and objects in the vicinity) monitor and apply rules of en-gagement to all applicable tracks in the
1127
+ local operating area
1128
+ 
1129
+  receive and send radio messagesto the battle group commander and other operating units in the area
1130
+  monitor tracks on the Aegis dis-play system and maintain situation
1131
+ awareness for all contacts of
1132
+ interest monitor performance of actionstaken by team members to assess the situation
1133
+  monitor the tactical action offi-cer's/commanding officer's performance maintain communications withCIC team members regarding their assessment of tracks and vari-ous actions taken
1134
+ In broad perspective, although teammembers spend much of their time in rou-
1135
+ tine activities, a number of different atten-
1136
+ tionally demanding, knowledge-intensive,
1137
+ and procedurally complex tasks may de-
1138
+ mand attention at any moment. Each of
1139
+ these tasks is usually triggered by a
1140
+ stimulus event, such as a communication
1141
+ from a team member or an alert, and, in
1142
+ order to obtain proper interpretation, mayrequire additional information-seeking be-havior. The cognitive challenge of select-
1143
+ ing and interpreting information to main-
1144
+ tain and revise one's SA is inherently
1145
+ complex. (Jager, Tenney, and Pew, 1995
1146
+ HF) Problems arise when, in the dynamic
1147
+ and multidimensional environments of
1148
+ some littoral antiair warfare scenarios, the
1149
+ situation-critical data become more time-
1150
+ compressed or ambiguous than humans
1151
+ can handle within the inherent time con-
1152
+ straints of the evolving scenario.Resources are such things as processing effort,the various forms of memory capacity, and
1153
+ communications channels (Bobrow & Norman,
1154
+ 1975, in Rasmussen et al).Topics to be discussed include: (1) a descriptionof the difficult tasks identified for analysis; (2) the
1155
+ general methodological approach; (3) develop-
1156
+ ment of the performance measures and issues
1157
+ related to their development; (4) discussion of the
1158
+ ; and (5) discussion of the types of errors made by
1159
+ decision makers and interpretations for the cause
1160
+ of these errors based in the cognitive psychology
1161
+ literature.
1162
+ Paper presented to the Second International Command and Control Research and Technology Symposium,20Monterey, Ca.tactical operations require decision makingconditions of time pressure, stress, am-
1163
+ biguous, inaccurate and missing informa-
1164
+ tion, uncertain communications, and
1165
+ shifting conditions. These conditions make
1166
+ it difficult to perform careful analysis prior
1167
+ to making decisions. Traditionally, most
1168
+ decision research and decision support
1169
+ system development have focused on well-
1170
+ defined tasks in carefully controlled envi-
1171
+ ronments. Recent research has suggested
1172
+ that tactical decision makers use experi-
1173
+ ence to generate a likely course of action
1174
+ and then evaluate its feasibility using
1175
+ mental simulation.One way to prevent the same type of casualtyfrom being repeated is to thoroughly investigate
1176
+ and analyze root causes of actual mishaps, as well
1177
+ as data collected in a simulated tactical
1178
+ environment, and apply the findings in a concrete
1179
+ manner to improve tactical decision making.According to the recognition-primed decision-making (RPD) model, experienced decision
1180
+ makers can make rapid, high-quality decisions by
1181
+ associating a situation directly with the actions
1182
+ that normally work well in that kind of situation.
1183
+ Roberts, K. H., Stout, S. K., and Halpern, J. J.(1994) Decision Dynamics in Two High
1184
+ Reliability Military Organizations. Man-agement Science, Vol. 40, No. 5, May1994, 614-624.Tetlock, P. E., Accountability and the Persever-ance of First Impressions, Social Psycho-logical Quarterly, 26(1983), 285-292.Tetlock, P. E., Accountability: The NeglectedSocial Context of Judgment and Choice, in
1185
+ L. L. Cummings and B. M. Straw (Eds.),
1186
+ Research in Organizational Behavior, Vol.7, JAI Press, Greenwich, CT, 1985, 297-332.The comparison to norms tool is based on acognitive model of human information processing
1187
+ which uses a feature matching strategy. The
1188
+ model proposes a data-driven process. As such,
1189
+ the comparison to norms tool is a knowledge-
1190
+ based tool with the knowledge represented as
1191
+ templates. Each template is a linked timeline
1192
+ associating individual features, through a series of
1193
+ feature matches, with expected actions. This
1194
+ allows the tool to make an assessment of the
1195
+ situation presented to the decision maker. A
1196
+ related module is the response manager. This
1197
+ assessment consists of a categorization of the
1198
+ situation (e.g., “hostile aircraft attacking”) and
1199
+ presentation of an associated template. The
1200
+ response manager module shows a template for an
1201
+ assessment of the current situation. This is a
1202
+ timeline display for template features, or events.
1203
+ The response manager module also provide inputs
1204
+ to the Prioritized track list, Alerts, and Responses
1205
+ and Tripwires.SABER is a model of another cognitivestrategy employed in making decisions. This
1206
+ strategy is known as explanation-based reasoning,
1207
+ or story generation. In this approach, available
1208
+ data are assembled into explanatory structures,
1209
+ with one structure for each possible conclusion.
1210
+ Each of the explanations attempts to explain how
1211
+ every piece of data can be accounted for in
1212
+ support of each conclusion, even though some of
1213
+ the data items would naturally contradict reaching
1214
+ some conclusions. Contradictory data are
1215
+ explained through the of internal assumptions. It
1216
+ is assumed that there are a fixed number of pre-
1217
+ defined possible conclusions and each data item
1218
+ points directly to one of those possible conclu-
1219
+ sions.Once the explanations are constructed,SABER evaluates them to determine which seems
1220
+ most plausible. Plausibility is based on three
1221
+ Paper presented to the Second International Command and Control Research and Technology Symposium,21Monterey, Ca.criteria: simplicity, completeness, and impor-tance. SABER provide data products which relate
1222
+ data to explanatory hypotheses for the current
1223
+ situation. Hypotheses are presented in rank order,
1224
+ with evidence for and against each hypothesis and
1225
+ missing data is presented below the corresponding
1226
+ hypothesis. For a more detailed description of the
1227
+ DSS the reader is referred to Hair & Pickslay,
1228
+ 1992; Hair, Pickslay, & Chow, 1992; Hutchins
1229
+ and Rummel, 1994.Hair, D. C. and Pickslay, K. (1992). Explanation-Based Reasoning in Decision Support
1230
+ Systems, Proceedings of the 9th AnnualConference on Command and Control De-cision Aids, June 1992, Monterey, CA.Hair, D. C., Pickslay, K. & Chow, S. (1992)Explanation-Based Decision Support in
1231
+ Real Time Situations. Proceedings of the1992 IEEE International Conference onTools with AI. Nov. 1992, Arlington, VA.Various studies have indicated that as much as 90
1232
+ percent of industrial and system failures are
1233
+ produced by human error (Senders & Moray,
1234
+ 1991).
1235
+ At the same time, a shift was occurring in U.S.
1236
+ Navy doctrine, away from a "blue-water" strategy
1237
+ to a doctrine of littoral operations aimed at
1238
+ potentially hostile regional powers.Tactical decision makers in today's oper-ating environment are required to perform
1239
+ complex tasks in a highly dynamic envi-
1240
+ ronment. Numerous interactive surface
1241
+ units and aircraft whose parameters are in
1242
+ flux and which must be continually
1243
+ sensed, processed, their future status pro-
1244
+ jected (development of hypotheses re-
1245
+ garding their future behavior) and actions
1246
+ taken to assure the successful outcome.Operating close to land presents additionalchallenges to the tactical decision maker.
1247
+ In littoral (i.e., near-land) settings, which
1248
+ most likely represent the majority of future
1249
+ anticipated naval conflicts, the decisions to
1250
+ be made are even more complex than they
1251
+ would be in full-scale warfare. When ci-
1252
+ vilian and neutral nation resources are in
1253
+ the conflict area incoming information
1254
+ carries an added element of uncertainty. In
1255
+ these situations, interpretation of the rules
1256
+ of engagement, contact identification, de-
1257
+ termining the capability and possible in-
1258
+ tent of the potential threat, and the
1259
+ shoot/no-shoot decision often pose ex-
1260
+ tremely difficult decision problems. (Since
1261
+ 70 percent of the world's population lives
1262
+ within 200 miles of the sea, most future
1263
+ contingencies are likely to involve littoral
1264
+ warfare (Mundy, 1994).)as opposed to replacing "the user's approach tothe problem" (emphasis in original, Cohen, 1993)with tools based on decision analysis or
1265
+ mathematical optimization,
1266
+ A considerably smaller number are attributable to
1267
+ other causes such as mechanical, electrical and
1268
+ materials failure (Meshkati, 1993). The purpose for this phase of the TADMUSprogram is to empirically evaluate the effective-
1269
+ ness of a DSS based on these recent approaches to
1270
+ decision support, to the extent possible,"This focus on why errors occuris...different from...the typical study of
1271
+ human errors which solely emphasize what
1272
+ occurs, a point of view which has received
1273
+ considerable criticism." (Rouse & Rouse,
1274
+ 1983, p. 539)
1275
+ Paper presented to the Second International Command and Control Research and Technology Symposium,22Monterey, Ca. DSS which will mitigate typical types oferrors. The objective during this phase of
1276
+ the research was to develop an under-
1277
+ standing of the decision-making problems
1278
+ presented by current and future Navy sce-
1279
+ narios in order to identify the types and
1280
+ forms of information that are likely to fa-
1281
+ cilitate performance of these activities.
1282
+ Rouse, W. B. & Valusek, J. (1993). Evo-lutionary Design of Systems to Support
1283
+ Decision Making. In G. A. Klein, J.
1284
+ Orasanu, R. Calderwood, & C. E.
1285
+ Zsambok (Eds.) Decision Making inAction: Models and Methods (pp.270-286). Ablex Publishing Corpora-
1286
+ tion, New Jersey.The Naturalistic Decision Making (NDM)model (Klein, 1989, 1993) seems more
1287
+ applicable than traditional decision-
1288
+ making models to the types of decisions
1289
+ involved in tactical decision-making.
1290
+ Early work conducted under the TADMUS
1291
+ program to determine the cognitive strate-
1292
+ gies employed by Navy AAW decision-
1293
+ makers found that when performing situa-
1294
+ tion assessment 87% of the time a recog-
1295
+ nitional strategy was used and 13% of the
1296
+ time story generation was used (Kaempf,
1297
+ Wolf, and Miller, 1993). We have applied
1298
+ these new models of human decision
1299
+ making—which parallel the cognitive
1300
+ strategies used by domain experts—to the
1301
+ design of a DSS for enhancing antiair war-
1302
+ fare tactical decision making. These two
1303
+ models for situation assessment are feature
1304
+ matching and story generation.
1305
+ (The scenarios were intentionally devel-
1306
+ oped to have many tracks with a high de-
1307
+ gree of uncertainty associated where it is
1308
+ not always clear whether a particular track
1309
+ should be engaged. Our interest in theTADMUS program was in gaining insightinto and aiding the decision process. The
1310
+ reader is referred to Hutchins, 1995, for a
1311
+ detailed coverage of performance meas-
1312
+ urement issues and scenario development
1313
+ issues.)
1314
+ These conditions include dynamic, fluid
1315
+ situations, time pressure, high risk, multi-
1316
+ ple decision makers, shifting and compet-
1317
+ ing goals, action-feedback loops, and
1318
+ situations with uncertain and incomplete
1319
+ data (Orasanu & Connolly, 1993).
adp1_02-1-376 ADDED
The diff for this file is too large to render. See raw diff
 
adp2_0-1-88 ADDED
The diff for this file is too large to render. See raw diff
 
adp3_0-1-102 ADDED
The diff for this file is too large to render. See raw diff
 
adp5_0-1-30 ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ DISTRIBUTION  RESTRICTION:  Approved  for  public  release;;  distribution  is  unlimited.HEADQUARTERS,  DEPARTMENT  OF  THE  ARMYMAY  2012THE  OPERATIONS  PROCESSADP  5-­0
2
+ This publication is available at Army Knowledge Online (https://armypubs.us.army.mil/doctrine/index.html).
3
+ *ADP 5-0 (FM 5-0) Army Doctrine Publication Headquarters Department of the Army No. 5-0 (FM 5-0) Washington, DC, 17 May 2012 The Operations Process Contents Page PREFACE ..................................................................................................... ii Definition and Purpose ........................................................................... 1 Principles of the Operations Process ..................................................... 2 Activities of the Operations Process....................................................... 6 Conclusion ........................................................................................... 16 GLOSSARY .................................................................................. Glossary-1 REFERENCES .......................................................................... References-1 Figures Figure 1. The operations process underlying logic ...............................................iv Figure 2. The operations process ......................................................................... 1 Table Table 1. Preparation activities ............................................................................ 11 DISTRIBUTION RESTRICTION. Approved for public release; distribution is unlimited. *This publication supersedes FM 5-0, dated 26 March 2010. i
4
+ Preface Army Doctrine Publication (ADP) 5-0, The Operations Process, constitutes the Army’s view on planning, preparing, executing, and assessing operations. (See figure 1 on page iv.) It accounts for the complex, ever-changing, and uncertain nature of operations and recognizes that a military operation is foremost a human undertaking. As such, this publication emphasizes the philosophy of mission command to include the central role of commanders (supported by their staffs) in driving the operations process. To comprehend the doctrine contained in ADP 5-0, readers must first understand the foundations of unified land operations described in ADP 3-0, Unified Land Operations. Readers must also fully understand the principles of mission command described in ADP 6-0, Mission Command. For a detailed explanation of the operations process, readers should refer to Army Doctrine Reference Publication (ADRP) 5-0, The Operations Process. The principal audience for ADP 5-0 includes Army commanders, leaders, and unit staffs (officers, noncommissioned officers, and Soldiers). Commanders and staffs of Army headquarters serving as a joint task force or multinational headquarters should also refer to applicable joint or multinational doctrine concerning the range of military operations as well as joint or multinational forces. Trainers and educators throughout the Army will also use this manual. Commanders, staffs, and subordinates ensure their decisions and actions comply with applicable U.S., international, and, in some cases, host nation laws and regulations. Commanders at all levels ensure their Soldiers operate in accordance with the law of war and the rules of engagement. (See Field Manual [FM] 27-10.) ADP 5-0 uses joint terms where applicable. Selected joint and Army terms and definitions appear in both the glossary and the text. Terms for which ADP 5-0 is the proponent publication (the authority) are marked with an asterisk (*) in the glossary. Definitions for which ADP 5-0 is the proponent publication are boldfaced in the text. These terms and their definitions will be in the next revision of FM 1-02. For other definitions shown in the text, the term is italicized and the number of the proponent publication follows the definition. ADP 5-0 applies to the Active Army, Army National Guard/Army National Guard of the United States, and United States Army Reserve unless otherwise stated. The proponent of ADP 5-0 is the United States Army Combined Arms Center. The preparing agency is the Combined Arms Doctrine Directorate, United States Army Combined Arms Center. Send comments and recommendations on a DA Form 2028 (Recommended Changes to Publications and Blank Forms) to Commander, U.S. Army Combined Arms Center and Fort Leavenworth, ATTN: ATZL-MCK-D (ADP 5-0), 300 McPherson Avenue, Fort Leavenworth, KS 66027-2337; by e-mail to usarmy.leavenworth.mccoe.mbx.cadd-org-mailbox@mail.mil; or submit an electronic DA Form 2028. ADP 5-0 17 May 2012 ii
5
+ ��������������������������������ACKNOWLEDGEMENT Cover photo courtesy of the U.S. Army at http://www.flickr.com/photos/soldiersmediacenter/6846045865/.
6
+ ������� iv ADP 5-0 17 May 2012 Figure 1. The operations process underlying logic
7
+ This publication defines and describes the operations process. It provides principles commanders and staffs consider to effectively plan, prepare, execute, and continuously assess operations. DEFINITION AND PURPOSE 1. The Army’s framework for exercising mission command is the operations process— the major mission command activities performed during operations: planning, preparing, executing, and continuously assessing the operation. Commanders, supported by their staffs, use the operations process to drive the conceptual and detailed planning necessary to understand, visualize, and describe their operational environment; make and articulate decisions; and direct, lead, and assess military operations. Figure 2. The operations process 2. The activities of the operations process are not discrete; they overlap and recur as circumstances demand. Planning starts an iteration of the operations process. Upon completion of the initial order, planning continues as leaders revise the plan based on changing circumstances. Preparing begins during planning and continues through execution. Execution puts a plan into action by applying combat power to seize, retain, and exploit the initiative to gain a position of relative advantage. Assessing is continuous and influences the other three activities. 3. Both the commander and staff have important roles within the operations process. The commander’s role is to drive the operations process as depicted in figure 2. The 17 May 2012 ADP 5-0 1
8
+ ADP 5-0 staff’s role is to assist commanders with understanding situations, making and implementing decisions, controlling operations, and assessing progress. In addition, the staff assists subordinate units (commanders and staffs), and keeps units and organizations outside the headquarters informed throughout the operations process. (ATTP 5-0.1 discusses the duties and responsibilities of the staff in detail.) PRINCIPLES OF THE OPERATIONS PROCESS 4. The philosophy of mission command guides commanders, staffs, and subordinates as they plan, prepare, execute, and assess operations. Mission command requires an environment of mutual trust and shared understanding among commanders, staffs, and subordinates. It requires a command climate in which commanders encourage subordinates to accept prudent risk and exercise disciplined initiative to seize opportunities and counter threats within the commander’s intent. Through mission orders, commanders focus their instructions on the purpose of the operation rather than on the details of how to perform assigned tasks. Doing this minimizes detailed control and allows subordinates the greatest possible freedom of action. Finally, when delegating authority to subordinates, commanders set the necessary conditions for success by allocating appropriate resources to subordinates based on assigned tasks. 5. Commanders and staffs use the operations process to integrate numerous tasks that are executed throughout the headquarters and with subordinate units. Commanders must organize and train their staffs and subordinates as an integrated team to simultaneously plan, prepare, execute, and assess operations. In addition to the principles of mission command discussed in ADP 6-0, commanders and staffs consider the following principles for the effective use of the operations process: �Commanders drive the operations process. �Build and maintain situational understanding. �Apply critical and creative thinking. �Encourage collaboration and dialogue. COMMANDERS DRIVE THE OPERATIONS PROCESS 6. Commanders are the most important participants in the operations process. While staffs perform essential functions that amplify the effectiveness of operations, commanders drive the operations process through understanding, visualizing, describing, directing, leading, and assessing operations. Understand 7. To understand something is to grasp its nature and significance. Understanding includes establishing context—the set of circumstances that surround a particular event or situation. Throughout the operations process, commanders develop and improve their understanding of their operational environment and the problem. An operational environment is a composite of the conditions, circumstances, and influences that affect the employment of capabilities and bear on the decisions of the commander (JP 3-0). Both conceptual and detailed planning assist commanders in developing their initial ADP 5-0 17 May 2012 2
9
+ The Operations Process understanding of the operational environment and the problem. Based on personal observations and inputs from others (to include running estimates), commanders improve their understanding and modify their visualization throughout the conduct of operations. Visualize 8. As commanders begin to understand their operational environment and the problem, they start visualizing a desired end state and potential solutions to solve the problem. Collectively, this is known as commander's visualization—the mental process of developing situational understanding, determining a desired end state, and envisioning an operational approach by which the force will achieve that end state. Commander’s visualization begins in planning and continues throughout the operations process until the force accomplishes the mission. During planning, commander’s visualization provides the basis for developing plans and orders. During execution, it helps commanders determine if, when, and what to decide, as they adapt to changing conditions. Describe 9. After commanders visualize an operation, they describe it to their staffs and subordinates to facilitate shared understanding and purpose. During planning, commanders ensure subordinates understand their visualization well enough to begin course of action development. During execution, commanders describe modifications to their visualization resulting in fragmentary orders that adjust the original order. Commanders describe their visualization in doctrinal terms, refining and clarifying it as circumstances require. Commanders express their visualization in terms of— �Commander’s intent. �Planning guidance, including an operational approach. �Commander’s critical information requirements. �Essential elements of friendly information. 10. The commander’s intent is a clear and concise expression of the purpose of the operation and the desired military end state that supports mission command, provides focus to the staff, and helps subordinate and supporting commanders act to achieve the commander’s desired results without further orders, even when the operation does not unfold as planned (JP 3-0). During planning, the initial commander's intent drives course of action development. In execution, the commander’s intent guides disciplined initiative as subordinates make decisions when facing unforeseen opportunities or countering threats. 11. In addition to issuing their commander’s intent, commanders provide planning guidance that conveys the essence of their visualization. Effective planning guidance broadly describes when, where, and how the commander intends to employ combat power to accomplish the mission within the higher commander’s intent. Planning guidance includes an operational approach—a description of the broad actions the force must take to transform current conditions into those desired at end state (JP 5-0). The 17 May 2012 ADP 5-0 3
10
+ ADP 5-0 operational approach forms the basis of the unit’s concept of operations and serves as the link between conceptual and detailed planning. 12. Commanders also describe gaps in their visualization by stating their commander’s critical information requirements (CCIRs). Commanders use CCIRs to focus information collection on the relevant information they need to make critical decisions throughout the conduct of operations. The two components of CCIRs are friendly force information requirements and priority intelligence requirements. 13. In addition to information commanders need, commanders also describe the information they want protected as essential elements of friendly information (EEFIs). EEFIs establish an element of information to protect rather than one to collect. EEFIs identify those elements of friendly force information that, if compromised, would jeopardize mission success. Direct 14. Commanders direct all aspects of operations by establishing their commander’s intent, setting achievable objectives, and issuing clear tasks to subordinate units. Throughout the operations process, commanders direct forces by— �Preparing and approving plans and orders. �Establishing command and support relationships. �Assigning and adjusting tasks, control measures, and task organization. �Positioning units to maximize combat power. �Positioning key leaders at critical places and times to ensure supervision. �Allocating resources to exploit opportunities and counter threats. �Committing the reserve as required. Lead 15. Through leadership, commanders provide purpose, direction, and motivation to subordinate commanders, their staff, and Soldiers. In many instances, a commander’s physical presence is necessary to lead effectively. Where the commander locates within the area of operations is an important leadership consideration. Commanders balance their time between leading the staff through the operations process and providing purpose, direction, and motivation to subordinate commanders and Soldiers away from the command post. Assess 16. Commanders continuously assess the situation to better understand current conditions and determine how the operation is progressing. Continuous assessment helps commanders anticipate and adapt the force to changing circumstances. Commanders incorporate the assessments of the staff, subordinate commanders, and unified action partners into their personal assessment of the situation. Based on their assessment, commanders modify plans and orders to adapt the force to changing circumstances. ADP 5-0 17 May 2012 4
11
+ The Operations Process BUILD AND MAINTAIN SITUATIONAL UNDERSTANDING 17. Success in operations demands timely and effective decisions based on applying judgment to available information and knowledge. As such, commanders and staffs seek to build and maintain situational understanding throughout the operations process. Situational understanding is the product of applying analysis and judgment to relevant information to determine the relationships among the operational and mission variables to facilitate decisionmaking. Building and maintaining situational understanding is essential for establishing the situation’s context, developing effective plans, assessing operations, and making quality decisions throughout the operations process. Commanders continually strive to maintain their situational understanding and work through periods of reduced understanding as the situation evolves. 18. Commanders and staffs use the operational and mission variables to help build their situational understanding. They analyze and describe an operational environment in terms of eight interrelated operational variables: political, military, economic, social, information, infrastructure, physical environment, and time (PMESII-PT). Upon receipt of a mission, commanders filter information categorized by the operational variables into relevant information with respect to the mission. They use the mission variables, in combination with the operational variables, to refine their understanding of the situation and to visualize, describe, and direct operations. The mission variables are mission, enemy, terrain and weather, troops and support available, time available, and civil considerations (METT-TC). APPLY CRITICAL AND CREATIVE THINKING 19. Commanders and staffs apply critical and creative thinking throughout the operations process to assist them with understanding situations, making decisions, and directing action. Critical thinking is purposeful and reflective judgment about what to believe or what to do in response to observations, experience, verbal or written expressions, or arguments. Creative thinking involves creating something new or original. Creative thinking leads to new insights, novel approaches, fresh perspectives, and new ways of understanding and conceiving things. 20. Critical and creative thinking are indispensible to the operations process. For both commanders and staff, these two skills begin with a rigorous analysis of friendly and enemy forces, as they relate to one another in time and space. This analysis includes weapons system ranges, mobility options afforded by terrain and weather, operational reach, communications system range, sustainment, and other considerations of the operational and mission variables. Disciplined and focused analysis of the operational and mission variables, coupled with critical and creative thinking about the challenges and opportunities resulting from that analysis, is essential to developing a full appreciation of the range of alternatives available to accomplish assigned missions. ENCOURAGE COLLABORATION AND DIALOGUE 21. Throughout the operations process, commanders encourage continuous collaboration and dialogue among commanders, staffs, and unified action partners to create shared 17 May 2012 ADP 5-0 5
12
+ ADP 5-0 understanding and facilitate unity of effort. Collaboration is two or more people or organizations working together toward common goals by sharing knowledge and building consensus. Dialogue is a way to collaborate by involving the candid exchange of ideas or opinions among participants that encourages frank discussions in areas of disagreement. 22. Commanders, staffs, and unified action partners collaborate and dialogue actively, sharing and questioning information, perceptions, and ideas to better understand situations and make decisions. Collaboration and dialogue assist in developing shared understanding and purpose, building teams, and making rapid adjustments during execution. ACTIVITIES OF THE OPERATIONS PROCESS 23. The operations process consists of the major mission command activities: planning, preparing, executing, and assessing. PLANNING 24. Planning is the art and science of understanding a situation, envisioning a desired future, and laying out effective ways of bringing that future about. Army leaders plan to create a common vision among subordinate commanders, staffs, and unified action partners for the successful execution of operations. Planning results in a plan or order that communicates this vision and directs actions to synchronize forces in time, space, and purpose for achieving objectives and accomplishing missions. Integrated Planning 25. Planning consists of two separate, but closely related, components: a conceptual component and a detailed component. Conceptual planning involves understanding the operational environment and the problem, determining the operation’s end state, and visualizing an operational approach. Conceptual planning generally corresponds to operational art and is the focus of the commander with staff support. Detailed planning translates the broad operational approach into a complete and practical plan. Generally, detailed planning is associated with the science of operations including the synchronization of the forces in time, space, and purpose. Detailed planning works out the scheduling, coordination, or technical problems involved with moving, sustaining, and synchronizing the actions of force as a whole toward a common goal. Effective planning requires the integration of both the conceptual and detailed components of planning. Planning and Operational Art 26. Operational art is the cognitive approach by commanders and staffs—supported by their skill, knowledge, experience, creativity, and judgment—to develop strategies, campaigns, and operations to organize and employ military forces by integrating ends, ways, and means (JP 3-0). Operational art guides the conceptual and detailed aspects of planning to produce executable plans and orders. Operational art applies to all aspects of ADP 5-0 17 May 2012 6
13
+ The Operations Process operations and integrates ends, ways, and means, while accounting for risk and opportunities, across the levels of war. 27. The elements of operational art (see ADRP 3-0) assist commanders and staffs in the application of operational art. These conceptual tools help commanders think through the challenges of understanding their operational environment, defining the problem, developing an operational approach, and articulating their planning guidance that drives more detailed planning. Army Planning Methodologies Elements of operational art • End state and conditions • Center of gravity • Decisive points • Lines of operations and lines of effort • Operational reach • Basing • Tempo • Phasing and transitions • Culmination • Risk 28. Army leaders employ three methodologies for planning. Commanders and staffs determine the appropriate mix of these methodologies based on the scope of the problem, their familiarity with it, the time available, and the availability of a staff. Methodologies that assist commanders and staffs with planning include— �Army design methodology. �Military decisionmaking process (MDMP). �Troop leading procedures (TLP). Army Design Methodology 29. The Army design methodology is a methodology for applying critical and creative thinking to understand, visualize, and describe unfamiliar problems and approaches to solving them. Army design methodology is an iterative process of understanding and problem framing that uses elements of operational art to conceive and construct an operational approach to solve identified problems. Commanders and their staffs use Army design methodology to assist them with the conceptual aspects of planning. 30. Army design methodology entails framing the operational environment, framing the problem, and developing an operational approach to solve the problem. Army design methodology results in an improved understanding of the operational environment, a problem statement, an initial commander’s intent, and an operational approach that serves as the link between conceptual and detailed planning. Based on their understanding and learning gained during Army design methodology, commanders issue planning guidance, to include an operational approach, to guide more detailed planning using the MDMP. 31. The understanding developed through Army design methodology continues through preparation and execution in the form of continuous assessment. Assessment, to include 17 May 2012 ADP 5-0 7
14
+ ADP 5-0 updated running estimates, helps commanders measure the overall effectiveness of employing forces and capabilities to ensure that the operational approach remains feasible and acceptable within the context of the higher commander’s intent and concept of operations. If the current operational approach fails to meet these criteria, or if aspects of the operational environment or problem change significantly, the commander may decide to reframe. Reframing involves revisiting earlier hypotheses, conclusions, and decisions that underpin the current operational approach. Reframing can lead to a new problem statement and operational approach, resulting in an entirely new plan. Military Decisionmaking Process 32. The military decisionmaking process is an iterative planning methodology to understand the situation and mission, develop a course of action, and produce an operation plan or order. The MDMP combines the conceptual and detailed aspects of planning and integrates the activities of the commander, staff, subordinate headquarters, and other partners throughout the planning process. The MDMP helps leaders apply thoroughness, clarity, sound judgment, logic, and professional knowledge to understand situations, develop options to solve problems, and reach decisions. The MDMP results in an improved understanding of the situation and a plan or order that guides the force through preparation and execution. 33. The MDMP facilitates collaborative and parallel planning as the higher headquarters solicits input and continually shares information concerning future operations with subordinate and adjacent units, supporting and supported units, and unified action partners through planning meetings, warning orders, and other means. Commanders encourage active collaboration among all organizations affected by the pending operations to build shared understanding, participate in course of action development and decisionmaking, and resolve conflicts before publication of the plan or order. 34. The MDMP consists of a series of steps that have various inputs and outputs. The outputs lead to an increased understanding of the situation facilitating the next step of the MDMP. Commanders and staffs generally perform these steps sequentially; however, they may revisit several steps in an iterative fashion, as they learn more about the situation before producing the plan or order. The steps of the MDMP are— �Step 1 – Receipt of mission. �Step 2 – Mission analysis. �Step 3 – Course of action development. �Step 4 – Course of action analysis. �Step 5 – Course of action comparison. �Step 6 – Course of action approval. �Step 7 – Orders production, dissemination, and transition. Troop Leading Procedures 35. Troop leading procedures are a dynamic process used by small-unit leaders to analyze a mission, develop a plan, and prepare for an operation. TLP are used by commanders and leaders without a staff. These procedures enable leaders to maximize ADP 5-0 17 May 2012 8
15
+ The Operations Process available planning time while developing effective plans and preparing their units for an operation. Like the MDMP, troop leading procedures consist of a series of steps: �Step 1 – Receive the mission. �Step 2 – Issue a warning order. �Step 3 – Make a tentative plan. �Step 4 – Initiate movement. �Step 5 – Conduct reconnaissance. �Step 6 – Complete the plan. �Step 7 – Issue the order. �Step 8 – Supervise and refine the plan. 36. The sequence of the steps of troop leading procedures is not rigid. Leaders modify them as required. Higher headquarters issue frequent warning orders to optimize available time for subordinates to conduct their TLP. Guides to Effective Planning 37. Planning helps commanders understand and develop solutions to problems, anticipate events, adapt to changing circumstances, task-organize the force, and prioritize efforts. Effective planning requires dedication, study, and practice. Planners must be technically and tactically competent within their areas of expertise and disciplined in the use of doctrinally correct terms and symbols. The following guides aid in effective planning: �Commanders focus planning. �Develop simple, flexible plans through mission orders. �Optimize available planning time. �Continually refine the plan. Commanders Focus Planning 38. Commanders are the most important participants in effective planning. They focus the planning effort by providing their commander’s intent, issuing planning guidance, and making decisions throughout the planning process. Commanders apply discipline to the planning process to meet the requirements of time, planning horizons, simplicity, level of detail, and desired outcomes. Commanders ensure that all operation plans and orders comply with applicable domestic and international laws. They also confirm that the plan or order is relevant and suitable for subordinates. Generally, the more involved commanders are in planning, the faster staffs can plan. Through personal involvement, commanders ensure the plan reflects their commander’s intent. Develop Simple, Flexible Plans Through Mission Orders 39. Effective plans and orders are simple and direct. Staffs prepare clear, concise orders that communicate an understanding of the operation through the use of doctrinally correct operational terms and symbols. Doing this minimizes chances of 17 May 2012 ADP 5-0 9
16
+ ADP 5-0 misunderstanding. Clarity and brevity are important. Shorter, rather than longer, plans aid in simplicity. Shorter plans are easier to disseminate, read, and remember. 40. Flexible plans help units adapt quickly to changing circumstances. Commanders and planners build opportunities for initiative into plans by anticipating events that allow them to operate inside of the enemy’s decision cycle or to react promptly to deteriorating situations. Identifying decision points and designing branches ahead of time—combined with a clear commander’s intent—help create flexible plans. 41. Commanders stress the importance of using mission orders as a way of building simple, flexible plans. Mission orders are directives that emphasize to subordinates the results to be attained, not how they are to achieve them (ADP 6-0). Mission orders clearly convey the unit’s mission and the commander’s intent. Mission orders focus subordinates on what to do and the purpose of doing it, without prescribing exactly how to do it. Commanders establish control measures to aid cooperation among forces without imposing needless restriction on freedom of action. Optimize Available Planning Time 42. Time is a critical variable in operations. Therefore, time management is important in planning. Whether done deliberately or rapidly, all planning requires the skillful use of available time to optimize planning and preparation throughout the unit. Taking more time to plan often results in greater synchronization; however, any delay in execution risks yielding the initiative—with more time to prepare and act—to the enemy. When allocating planning time to the staff, commanders must ensure subordinates have enough time to plan and prepare their own actions prior to execution. Commanders follow the “one-third—two-thirds rule” as a guide to allocate time available. They use one-third of the time available before execution for their planning and allocate the remaining two-thirds of the time available before execution to their subordinates for planning and preparation. Continually Refine the Plan 43. Planning does not cease with production of a plan or order. It continues throughout an operation as the order is refined based on confirmation briefings, rehearsals and changes in the situation. In addition, staffs are always refining plans for branches and sequels throughout an operation. During preparation and execution, the plan is continuously refined as situational understanding improves. Through assessment, subordinates and others provide feedback on the progress of operations. In some circumstances, commanders may determine that the current order (to include associated branches and sequels) is no longer relevant to the situation. In these instances, instead of modifying the current plan, commanders reframe the problem and develop an entirely new plan. PREPARING 44. Preparation consists of those activities performed by units and Soldiers to improve their ability to execute an operation. Preparation creates conditions that improve friendly forces’ opportunities for success. It requires commander, staff, unit, and 10 ADP 5-0 17 May 2012
17
+ The Operations Process Soldier actions to ensure the force is trained, equipped, and ready to execute operations. Effective preparation helps commanders, staffs, and subordinate units better understand the situation and their roles in upcoming operations. The major activities of preparation are listed in table 1. Table 1. Preparation activities Continue to coordinate and conduct liaison Conduct rehearsals Initiate information collection Conduct plans-to-operations transitions Initiate security operations Revise and refine the plan Initiate troop movement Integrate new Soldiers and units Initiate sustainment preparations Complete task organization Initiate network preparations Train Manage terrain Perform pre-operations checks and inspections Prepare terrain Continue to build partnerships and teams Conduct confirmation briefs 45. Mission success depends as much on preparation as on planning. Higher headquarters may develop the best of plans; however, plans serve little purpose if subordinates do not receive them in time. Subordinates need enough time to fully comprehend the plan, rehearse key portions of the plan, and ensure Soldiers and equipment are positioned and ready to execute the operation. The following guidelines aid in effective preparation: �Secure and protect the force. �Improve situational understanding. �Understand, rehearse, and refine the plan. �Integrate, organize, and configure the force. �Ensure forces and resources are ready and positioned. Secure and Protect the Force 46. The force as a whole is often most vulnerable to surprise and enemy attack during preparation. As such, security operations—screen, guard, cover, area security, and local security—are essential during preparation. In addition, commanders ensure the various tasks of the protection warfighting function are fully integrated to safeguard bases, secure routes, and protect the force, while it prepares for operations. Improve Situational Understanding 47. During preparation, commanders may realize that their initial understanding developed during planning may be neither accurate nor complete. As such, commanders strive to validate assumptions and improve their situational understanding as they prepare for operations. Information collection (to include reconnaissance, surveillance, and intelligence operations) helps improve understanding of the enemy, terrain, and civil 17 May 2012 ADP 5-0 11
18
+ ADP 5-0 considerations. Inspections, rehearsals, liaison, and coordination help leaders improve their understanding of the friendly force. Understand and Rehearse the Plan 48. A successful transition from planning to execution requires those charged with executing the order to understand the plan fully. The transition between planning and execution takes place both internally in the headquarters and externally between the commander and subordinate commanders. Rehearsals, to include confirmation briefings and plans-to-operations transition briefings, help improve understanding of the concept of operations, control measures, decision points, and command and support relationships. Integrate, Organize, and Configure the Force 49. During preparation, commanders allocate time to put the new task organization into effect. This includes detaching units, moving forces, and receiving and integrating new units and Soldiers into the force. When units change task organization, they need preparation time to learn the gaining unit’s standard operating procedures and the plan the gaining unit will execute. The gaining unit needs preparation time to assess the new unit’s capabilities and limitations and to integrate new capabilities. Ensure Forces and Resources are Ready and Positioned 50. Effective preparation ensures that the right forces are in the right place, at the right time, with the right equipment and other resources ready to execute the operation. Concurrent with task organization, commanders use troop movement to position or reposition forces to the correct locations prior to execution. This includes positioning sustainment units and supplies. EXECUTING 51. Execution is putting a plan into action by applying combat power to accomplish the mission. During execution, commanders, staffs, and subordinate commanders focus their efforts on translating decisions into actions. They apply combat power to seize, retain, and exploit the initiative to gain and maintain a position of relative advantage. This is the essence of unified land operations (see ADP 3-0). Decisionmaking During Execution 52. Decisionmaking is tied to disciplined initiative and inherent in executing operations. Commanders observe the progress of operations and intervene when necessary to ensure success. Because operations never unfold exactly as envisioned and because understanding of the situation changes, a commander’s decisions made during execution are critical to an operation’s success. During execution, commanders direct their units forcefully and promptly to overcome the difficulties of enemy action, friendly errors, and other changes in their operational environment. 53. Commanders make execution and adjustment decisions throughout execution. Execution decisions implement a planned action under circumstances anticipated in the 12 ADP 5-0 17 May 2012
19
+ The Operations Process order. An execution decision is normally tied to a decision point—a point in space or time the commander or staff anticipates making a key decision concerning a specific course of action (JP 5-0). An adjustment decision is the selection of a course of action that modifies the order to respond to unanticipated opportunities or threats. An adjustment decision may include a decision to reframe the problem and develop an entirely new plan. 54. Executing, adjusting, or abandoning the original operation is part of decisionmaking in execution. By fighting the enemy and not the plan, successful commanders balance the tendency to abandon a well-conceived plan too soon against persisting in a failing effort too long. Effective decisionmaking during execution— �Relates all actions to the commander’s intent and concept of operations to ensure they support the decisive operation. �Is comprehensive, maintaining integration of combined arms rather than dealing with separate functions. �Relies heavily on intuitive decisionmaking by commanders and staffs to make rapid adjustments. �Is continuous and responds effectively to any opportunity or threat. Guides to Effective Execution 55. During execution, the situation may change rapidly. Operations the commander envisioned in the plan may bear little resemblance to actual events in execution. Subordinate commanders need maximum latitude to take advantage of situations and meet the higher commander’s intent when the original order no longer applies. Effective execution requires leaders trained and educated in independent decisionmaking, aggressiveness, and risk taking in an environment of mission command. During execution, leaders must be able and willing to solve problems within the commander’s intent without constantly referring to higher headquarters. Subordinates need not wait for top-down synchronization to act. The following guides aid in effective execution: �Seize the initiative through action. �Accept prudent risk to exploit opportunities. Seize the Initiative Through Action 56. Commanders create conditions for seizing the initiative by acting. Without action, seizing the initiative is impossible. Faced with an uncertain situation, people naturally tend to hesitate and gather more information to reduce the uncertainty. Although waiting and gathering information might reduce uncertainty, it will not eliminate it. Waiting may even increase uncertainty by providing the enemy with time to seize the initiative. It is far better to manage uncertainty by acting and developing the situation. Accept Prudent Risk to Exploit Opportunities 57. Uncertainty and risk are inherent in all military operations. Successful commanders are comfortable operating under conditions of uncertainty, as they balance various risks while taking advantage of opportunities. Prudent risk is a deliberate exposure to potential 17 May 2012 ADP 5-0 13
20
+ ADP 5-0 injury or loss when the commander judges the outcome in terms of mission accomplishment as worth the cost (ADP 6-0). Reasonably estimating and intentionally accepting risk is not gambling. Gambling, in contrast to taking prudent risk, is staking the success of an entire action on a single event without considering the hazard to the force should the event not unfold as envisioned. Therefore, commanders avoid taking a gamble. Commanders carefully determine risks, analyze and minimize as many hazards as possible, and then take prudent risks to exploit opportunities. ASSESSING 58. Assessment is the determination of the progress toward accomplishing a task, creating an effect, or achieving an objective (JP 3-0). Assessment is a continuous activity of the operations process. The focus of assessment, however, changes for each operations process activity. During planning, assessment focuses on understanding current conditions of an operational environment and developing an assessment plan, including what and how to assess progress. During preparation, assessment focuses on determining the friendly force’s readiness to execute the operation and on verifying the assumptions on which the plan is based. During execution, assessment focuses on evaluating progress of the operation. Based on their assessment, commanders direct adjustments to the order, ensuring the operation stays focused on accomplishing the mission. Assessment Process 59. Assessment involves the continuous monitoring and evaluation of the current situation to determine progress of an operation. Broadly, assessment consist of the following activities: �Monitoring the current situation to collect relevant information. �Evaluating progress toward attaining end state conditions, achieving objectives, and completing tasks. �Recommending or directing action for improvement. 60. Primary tools for assessing include running estimates, after action reviews, and the assessment plan. Running estimates provide information, conclusions, and recommendations from the perspective of each staff section. Running estimates help to refine the common operational picture and supplement it with information not readily displayed. Both formal and informal after action reviews help identify what was supposed to happen, what went right, and what went wrong for a particular action or operation, and how the commander and staff should do things differently in the future. The assessment plan includes measures of effectiveness, measures of performance, and indicators that help the commander and staff evaluate progress toward accomplishing tasks and achieve objectives. (See ATTP 5-01.1 for doctrine on building assessment plans.) Running Estimates 61. Effective plans and successful preparation, execution and assessment hinge on accurate running estimates. A running estimate is the continuous assessment of the 14 ADP 5-0 17 May 2012
21
+ The Operations Process current situation used to determine if the current operation is proceeding according to the commander’s intent and if planned future operations are supportable. Detailed running estimates begin in planning and are continuously updated during preparation and execution. In their running estimates, the command and staff continuously consider the effects of new information and update facts, assumptions, and conclusions. Running estimates from the staff always include recommendations to the commander. Guides to Effective Assessment 62. Throughout the conduct of operations, commanders integrate their own assessments with those of the staff, subordinate commanders, and other partners in the area of operations. The following guides aid in effective assessment: �Commanders prioritize the assessment effort. �Incorporate the logic of the plan. �Use caution when establishing cause and effect. �Combine quantitative and qualitative indicators. Commanders Prioritize the Assessment Effort 63. Commanders establish priorities for assessment in their planning guidance, CCIRs, and decision points. By prioritizing the effort, commanders avoid excessive analyses when assessing operations. Committing valuable time and energy to developing excessive and time-consuming assessment schemes squanders resources better devoted to other operations process activities. Commanders reject the tendency to measure something just because it is measurable. Effective commanders avoid burdening subordinates and staffs with overly detailed assessments and collection tasks. Generally, the echelon at which a specific operation, task, or action is conducted should be the echelon at which it is assessed. Incorporate the Logic of the Plan 64. Effective assessment relies on an accurate understanding of the reasons and logic used to build the plan. Each plan is built on assumptions and an operational approach. The reasons and logic why the commander believes the plan will produce the desired results are important considerations when staffs determine how to assess operations. Recording and understanding this logic helps the staffs recommend the appropriate measures of effectiveness, measures of performance, and indicators for assessing the operation. Use Caution When Establishing Cause and Effect 65. Although establishing cause and effect is sometimes difficult, it is crucial to effective assessment. Sometimes, establishing causality between actions and their effects can be relatively straightforward, such as in observing a bomb destroy a bridge. In other instances, especially regarding changes in human behavior, attitudes, and perception, 17 May 2012 ADP 5-0 15
22
+ ADP 5-0 establishing links between cause and effect proves difficult. Commanders and staffs must guard against drawing erroneous conclusions in these instances. Combine Quantitative and Qualitative Indicators 66. Effective assessment incorporates both quantitative (observation-based) and qualitative (opinion-based) indicators. Human judgment is integral to assessment. A key aspect of any assessment is the degree to which it relies upon human judgment and the degree to which it relies upon direct observation and mathematical rigor. Rigor offsets the inevitable bias, while human judgment focuses rigor and processes on intangibles that are often key to success. The appropriate balance depends on the situation— particularly the nature of the operation and available resources for assessment—but rarely lies at the ends of the scale. CONCLUSION 67. The doctrine in this publication provides a starting point for the execution of the operations process. It establishes a common frame of reference and the intellectual tools Army leaders use to plan, prepare for, execute, and assess operations. By establishing a common approach and language, the doctrine in this publication promotes mutual understanding and enhances the effectiveness during the conduct of operations. The doctrine in this publication is a guide for action rather than a set of fixed rules. While it provides an authoritative guide for leaders and Soldiers, it requires original application adapted to circumstances. In operations, effective leaders recognize when and where doctrine, training, or even their past experience no long fits the situation and adapt accordingly. 16 ADP 5-0 17 May 2012
23
+ Glossary The glossary lists acronyms and terms with Army or joint definitions. Where Army and joint definitions differ, (Army) precedes the definition. Terms for which ADP 5-0 is the proponent are marked with an asterisk (*). The proponent publication for other terms is listed in parentheses after the definition. SECTION I – ACRONYMS AND ABBREVIATIONS ADP Army doctrine publication ADRP Army doctrine reference publication CCIR commander’s critical information requirement DA Department of the Army EEFI essential element of friendly information FM field manual JP joint publication MDMP military decisionmaking process METT-TC mission, enemy, terrain and weather, troops and support available, time available, and civil considerations PMESII-PT political, military, economic, social, information, infrastructure, physical environment, and time TLP troop leading procedures U.S. United States SECTION II – TERMS *Army design methodlogy A methodology for applying critical and creative thinking to understand, visualize, and describe unfamiliar problems and approaches to solving them. assessment The determination of the progress toward accomplishing a task, creating an effect, or achieving an objective. (JP 3-0) commander’s intent A clear and concise expression of the purpose of the operation and the desired military end state that supports mission command, provides focus to the staff, and helps subordinate and supporting commanders act to achieve the commander’s desire result withou furthor order, even when the operation does not unfold as planned. (JP 3-0) 17 May 2012 ADP 5-0 Glossary-1
24
+ Glossary *commander’s visualization The mental process of developing situational understanding, determining a desired end state, and envisioning an operational approach by which the force will achieve that end state. decision point A point in space or time the commander or staff anticipate making a key decision concerning a specific course of action. (JP 5-0) *execution Putting a plan into action by applying combat power to accomplish the mission. *military decisionmaking process An iterative planning methodology to understand the situation and mission, develop a course of action, and produce an operation plan or order. mission orders Directives that emphasize to subordinates the results to be attained, not how they are to achieve them. (ADP 6-0) operational approach A description of the broad action the force must take to transform current conditions into those desired at end state. (JP 5-0) operational art The cognitive approach by commanders and staffs—supported bytheir skill, knowledge, experience, creativity, and judment—to develop strategies, campaigns, and operations to organize and employ military forces by integrating ends, ways, and means. (JP 3-0) operational environment A composite of the conditions, circumstances, and influences that affect the employment of capabilities and bear on the decisions of the commander. (JP 3-0) *operations process The major mission command activities performed during operations: planning, preparing, executing, and continuously assessing the operation. *planning The art and science of understanding a situation, envisioning a desired future, and laying out effective ways of bringing that future about. *preparation Those activities performed by units and Soldiers to improve their ability to execute an operation. prudent risk A deliberate exposure to potential injury or loss when the commander judges the outcome in terms of mission accomplishment as worth the cost. (ADP 6-0) Glossary-2 ADP 5-0 17 May 2012
25
+ Glossary *running estimate The continuous assessment of the current situation used to determine if the current operation is proceeding according to the commander’s intent and if planned future operations are supportable. *situational understanding The product of applying analysis and judgment to relevant information to determine the relationships among the operational and mission variables to facilitate decisionmaking. *troop leading procedures A dynamic process used by small-unit leaders to analyze a mission, develop a plan, and prepare for an operation. 17 May 2012 ADP 5-0 Glossary-3
26
+ This page intentionally left blank.
27
+ References Field manuals and selected joint publications are listed by new number followed by old number. REQUIRED PUBLICATIONS These documents must be available to intended users of this publication. FM 1-02 (101-5-1). Operational Terms and Graphics. 21 September 2004. JP 1-02. Department of Defense Dictionary of Military and Associated Terms. 8 November 2010. RELATED PUBLICATIONS These documents contain relevant supplemental information. JOINT PUBLICATIONS Most joint publications are available online: <http://www.dtic.mil/doctrine/new_pubs/jointpub.htm.> JP 3-0. Joint Operations, 11 August 2011. JP 3-33. Joint Task Force Headquarters. 16 February 2007. JP 5-0. Joint Operation Planning, 11 August 2011. ARMY PUBLICATIONS Most Army doctrinal publications are available online: <http://www.apd.army.mil>. ADP 3-0 (FM 3-0). Unified Land Operations, 10 October 2011. ADP 6-0 (FM 6-0). Mission Command. 17 May 2012. ADRP 3-0. Unified Land Operations. 16 May 2012. ADRP 5-0. The Operations Process. 17 May 2012. ADRP 6-0. Mission Command. 17 May 2012. ATTP 5-0.1. Commander and Staff Officer Guide, 14 September 2011. FM 27-10. The Law of Land Warfare, 18 July 1956. WEB SITE Cover photo courtesy of the U.S. Army at http://www.flickr.com/photos/soldiersmediacenter/6846045865/. REFERENCED FORMS DA Form 2028. Recommended Changes to Publications and Blank Forms. 17 May 2012 ADP 5-0 References-1
28
+ This page intentionally left blank.
29
+ ADP 5-0 17 May 2012 By order of the Secretary of the Army: RAYMOND T. ODIERNO General, United States Army Chief of Staff Official: JOYCE E. MORROW Administrative Assistant to the Secretary of the Army 1211423 DISTRIBUTION: Active Army, Army National Guard, and U.S. Army Reserve: To be distributed in accordance with the initial distribution number (IDN) 110412, requirements for ADP 5-0.
30
+ PIN: 102805-000
fm1-1-132 ADDED
The diff for this file is too large to render. See raw diff
 
fm2-1-328 ADDED
The diff for this file is too large to render. See raw diff
 
fm3-1-280 ADDED
The diff for this file is too large to render. See raw diff
 
navrules-1-226 ADDED
The diff for this file is too large to render. See raw diff