Course Handout - Situation(al) Awareness (SA) in Effective Command and Control

Copyright Notice: This material was written and published in Wales by Derek J. Smith (Chartered Engineer). It forms part of a multifile e-learning resource, and subject only to acknowledging Derek J. Smith's rights under international copyright law to be identified as author may be freely downloaded and printed off in single complete copies solely for the purposes of private study and/or review. Commercial exploitation rights are reserved. The remote hyperlinks have been selected for the academic appropriacy of their contents; they were free of offensive and litigious content when selected, and will be periodically checked to have remained so. Copyright © 2018, Derek J. Smith.

 

First published online 08:28 GMT 17th March 2003, Copyright Derek J. Smith (Chartered Engineer). This version [2.1 - links to graphics] dated 09:00 BST 14th June 2018

 

 

 

1 - The Historical Background

 

"My experience is what I agree to attend to" (James, 1890/1950, pI.402).

 

The nineteenth century philosopher-psychologist William James once defined the difference between sensation and perception as resting in the fact that the latter included "the consciousness of farther facts" associated with the former (James, 1890/1950, pII.77). When you perceive something, James argued, you become aware not just of the existence of an object in the external world - its shape, colour, height, size, and speed, say - but also, thanks to your previous experience of the world, of that object's implications for you. Is it, for example, good to eat, or likely to eat you, or useful as a tool, or any one of a thousand other things; and what new behaviour on your part (if any) is warranted by its presence, and how urgently? As a result, your thoughts are punctuated every few moments by a new cycle of <object - implication> processing, and this has ramifications, in turn, for how we might explain the phenomenon of consciousness. In the event, James proposed a "personal consciousness" with five key characteristics, as follows .....

 

1. One's personal consciousness contains all one's thoughts.

 

2. These thoughts are ever-changing.

 

3. Within one's personal consciousness, thought is "sensibly continuous" (pI.237). That is to say, thought "feels" continuous even when we can point to objective time gaps. "Consciousness," James concluded, "does not appear to itself chopped up in bits" (pI.239).

 

4. One's personal consciousness "deals with objects independent of itself; that is, it is cognitive, or possesses the function of knowing" (pI.271).

 

5. One's personal consciousness, James wrote, is "interested in some parts of these objects to the exclusion of others, and welcomes or rejects - chooses from among them - all the while" (pI.225; italics original). In other words, it selectively attends.

 

It is James's fifth point which is most relevant to the present paper, because the ability to choose between inputs - that is to say, the mental faculty of "selective attention" - is central to our momentary understanding of the world. The objects it selects largely determine the subject matter of consciousness, and their implications largely determine the quality of the resulting experience. Moreover, it was in discussing selective attention that James introduced the related notion of the "span of consciousness", a concept which James defined as "the number of things we may attend to" (pI.405), and saw as being severely limited .....

 

ASIDE: James described a number of earlier studies which had tried to determine the maximum array size of briefly visually presented items whose number could be known instantly by an observer (that is to say, without the need for counting). He quotes data from Jevons (1871) to the effect that errors were rare on displays of up to five items, and then started to appear at six items (his subjects scored 120 out of a maximum 147 with six item arrays). Woodworth and Marquis (1949) call this limitation the "span of apprehension", define it as "the amount of material that an observer can grasp or 'apprehend' in a single act of perception" (p416), and place the point of total breakdown of accurate perception at "about 12" items. Past this point, observers can only "give an estimate".

 

Both of James's new topic areas were, of course, unremittingly cognitivist, meaning that neither received much scientific attention during the behaviourist interregnum. The study of selective attention then suddenly came back into fashion in the wake of a series of new age experiments carried out in the early 1950s (eg. Cherry, 1953). The study of consciousness took a little longer to become accepted, but is now the central topic of interest and investigation within the whole of cognitive science. Indeed, there are several distinct perspectives to its study, including .....

 

the philosophical perspective - this line of enquiry looks at what consciousness really is.

 

the computational perspective - this line of enquiry considers how to build artificially conscious machines.

 

the neuropsychological perspective (macro) - this line of enquiry looks for the brain centres responsible for consciousness.

 

the neuropsychological perspective (micro) - this line of enquiry looks at what is happening at neuronal level (or smaller) to deliver consciousness.

 

the clinical perspective - this line of enquiry looks at how consciousness-impairing diseases can be better treated.

 

the safety perspective - this line of enquiry looks at the limitations of consciousness behind the steering wheel (drivers, pilots, etc.), in the control room (power stations, etc.), or in the boardroom (disastrous decision making).

 

the forensic perspective - this line of enquiry looks at how said limitations can be diagnosed after a disaster .....

 

the design perspective - ..... and prevented from recurring by designing better systems.

 

the military perspective - this line of enquiry looks at how consciousness can be broadened so as to enhance individual or group battlefield performance.

 

The first of these perspectives has been with us for several thousand years, and, ultimately, cannot be resolved in isolation because it has no data, only opinion. The next three perspectives are much younger, have more empirical data than they can yet interpret, and are already individually massive supersciences. However, it is the last four perspectives which interest us here, for as a topic cluster they make up the science of "situation(al) awareness (SA)" .....

 

 

2 - Early Work

 

"Found difficulty in thinking of everything at once, and stalled too high in first attempt at landing" (Kenneth Craik, after trying out the Link Trainer; undated pre-1945; as cited in Sherwood, 1966, p115).

 

The idea that we have to struggle to control our passing through the world, especially when "at the wheel" in some way, is not new within psychology. During the 1930s, for example, Frederick (later Sir Frederick) Charles Bartlett (1886-1969), director of the experimental psychology laboratories at Cambridge University, carried out research on behalf of the RAF, studying amongst other things ways of reducing air accidents by better selection and training procedures (eg. Bartlett, 1937). This type of work automatically became more important during World War Two, and drew in a number of talented young postgraduates who would go on, after the war, to transform psychology. At Cambridge, this included Norman Mackworth and Kenneth Craik (1914-1945), and across the Atlantic it included Alphonse Chapanis (1917-2002) (the "father of ergonomics"), J.P.Guilford, Paul M. Fitts, and George A. Miller. Craik joined Bartlett's laboratories in 1936 as a doctoral student, and when war came became involved in researching the design of cockpit simulators (Bartlett, 1946). He played an important part in the development of the "Cambridge Cockpit", a model of good aircraft design, and suggested many practical improvements to both controls and instrumentation in recognition of the many physical, physiological, and perceptuo-motor limitations of the pilots being trained.

 

As for pilots' higher cognitive functions, it made no difference whether you focused on divided attention, selective attention, sustained attention, or decision making, the finding was always the same, namely that the cognitive system was routinely severely limited. Studies of divided attention found a cognitive system which could only divide itself so many times, studies of selective attention found a system which took time to select its target and was then easy to distract, studies of sustained attention found that even the fittest and most motivated personnel could only go so far in watching a screen or peering through a pair of binoculars before they started seeing things (or, worse, not seeing things), and studies of decision making found depressing levels of carelessness and poor judgement just about everywhere.

 

ASIDE: For a recent study of divided attention, see Wickens (1997). The sustained attention studies were amongst the first to enter civilian psychology after the war, as the psychology of "vigilance" (eg. Mackworth, 1950). The study of vigilance as a psychological construct has an abstract dated 1923 by no less an authority than the famous neurologist Sir Henry Head. However, Head saw vigilance as being much like general awareness, rather than the much narrower capacity for sustained perceptual performance under conditions of monotony. Many examples of ultimately disastrous decision making are given in our online resources in military, aerospace, railway, maritime, and IT project management disasters.

 

Thus far, the study of cockpit cognition had simply been an extension of the prewar science of "human factors", which was itself an offshoot from the tradition of "scientific management", the formal study of industrial and commercial processes which had taken shape in the US more than half a century beforehand. The lead author on this particular timeline was Frederick Winslow Taylor (1856-1915). It was Taylor, for example, who had introduced "time and motion study" methods into late nineteenth century industry, and who formalised what would nowadays be termed "evidence-based management". After Taylor came Charles Samuel Myers (1873-1946), who referred to the science of humankind at work as "industrial psychology" (Myers, 1929). In 1912 Myers founded the Cambridge psychology laboratories, and became its first Director (thus Bartlett's predecessor). The new science's own journal, Human Factors, was first published in the early 1930s. After 1945, the term "human engineering" became popular for a while in the US (Roscoe, 1995/2003 online), and "applied psychology" in the UK. There were further name changes and reorientations during the 1950s. Paul Fitts led the movement towards "engineering psychology" in 1951 (Fitts, 1951), but as that decade unfolded the name "ergonomics", the science of humankind at work (Greek: ergon = effort), grew more popular. The journal Ergonomics was launched in 1957. Finally, in the early 1960s, the concept of "cognitive engineering" was introduced (Bonjer, 1962), to be followed in 1977 by the term "cognitive ergonomics" (Newman, 1977).

 

ASIDE: For a fuller discussion of the modern names and angles of approach, see Stanton's (1996) special edition of The Psychologist or Lintern's (1999) special edition of The International Journal of Aviation Psychology.

 

 

3 - Motor Skills

Much of the following section appeared in Smith (1996; Chapter 1). It is repeated here with minor amendments and supported with hyperlinks.

The psychological investigation of motor skills began late in the nineteenth century as the processes of industrialisation began to place a premium on the rapid training of machine operators. Bryan and Harter (1897, 1899) investigated how quickly Morse code operators could learn the necessary keying and decoding skills. They interviewed and observed both experienced and trainee Western Union telegraph operators and plotted learning curves of performance against practice, finding distinct periods where continuing practice seemed to have no effect on ability. Because these periods showed up as flat areas of the learning curve, Bryan and Harter called them plateaus (optionally, plateaux). The topic was further investigated by William Book in the early years of the twentieth century. He studied the rate of acquisition of typewriting skills, and concluded as follows:

 

"There is in all [subjects] a rapid and continuous rise in the early stages of practice followed by a slower and more gradual rise in the later stages. All show marked daily variations [and] short periods of non-improvement." (Book, 1908, p158.)

 

In fact, Book distinguished two types of non-improvement. The first - called "breathing points" - lasted between six and eight days and were found in all subjects, and the second - the plateaus first noted by Bryan and Harter - lasted 17-33 days but were not found in all subjects. Other popular devices for testing motor skills were soon devised, including .....

 

The Pursuit Rotor: This task requires subjects to keep a stylus in contact with a target point mounted off-centre on a turntable. (The test exists in both one- and two-handed versions.)

 

Linear Tracking: This task requires subjects to keep a stylus in contact with a moving spot viewed through a slot.

 

Steering and Fine Control: This task requires subjects to control a piece of machinery like a bicycle, car, or aeroplane (or, at least, a simulator thereof). (Note that aircraft simulators and cars both involve simultaneous hand and foot activity.)

 

Mirror Drawing: This task requires subjects to trace round a line pattern viewed in a mirror.

 

Upside Down Letter Drawing: This task requires subjects to invert text letter-by-letter as quickly as they can working from right to left and up the page (so that the response sheet becomes readable as normal if simply rotated through 180o).

 

Track Tracing: This task requires subjects to move a looped electrical contact along a bent wire without touching it (a task now commonly used as a fete sideshow).

 

As to why plateaus should exist, Bryan and Harter argued that learning a skilled act involves setting up a hierarchy of habits. At the bottom of this hierarchy are discrete low-level motor acts such as transcribing or transmitting the individual Morse letters, then comes the ability to send firstly syllables and eventually whole words as known strings, and at the top of the hierarchy comes the ability to send firstly phrases and eventually whole sentences as known strings. Singer (1968) calls these first-, second-, and third-order habits respectively. Using tennis skills as his example, he identifies the lowest order habits as the basic strokes, the middle order habits as combinations of strokes and movements on the move, and the highest order habits as tactical game plans and the like.

 

As to why plateaus are important, the standard argument is that as learning takes place the brain reorganises itself so that more and more information can be processed at a time. It is just that it takes time and lots of practice for this mental reorganisation to take place. Plateaus occur whenever a lower-order habit approaches its absolute maximum speed, but cannot yet be invoked sufficiently automatically for concentration to be switched to developing the next higher-order habit.

 

It is also possible to identify more than one category of motor task, each making different demands on different types of memory, and each therefore following a differently shaped learning curve. Miller (1969) lists four possible categories of motor task, as follows:

 

Category I - Adjustive Tasks: This category of skill involves continuous perceptual-motor adjustment, and includes such tasks as tracking, pursuit, and steering.

 

Category II - Response Selection Tasks: This category of skill involves selection of appropriate responses from a repertoire, and includes such tasks as typing, sight reading of music, and sending Morse code.

 

Category III - Procedural Tasks: This category of skill involves complex, step-by-step, skills such as piloting an aircraft or assembling something from its components. (Thus making it a skill of executing other skills.)

 

Category IV - Skilled Performance Tasks: This category of skill involves complex, but nevertheless discrete, motor skills such as throwing and catching, ball control, and athletic skills such as vaulting and diving.

 

But the count could well be higher even than that. Following a series of factor analytic studies going back to 1954, Fleishman (1972) identified nine indepently variable physical components of human motor ability and eleven independently variable mental factors. He called the former physical proficiency factors, and the latter psychomotor ability factors. Here is what they seem to measure:

 

Physical Proficiency Factors: 1 - Extent Flexibility: This is the ability to flex-stretch the trunk or limbs to their limits in their various planes of movement. 2 - Dynamic Flexibility: This is the ability to make repeated rapid flex-stretchings of the trunk and limbs. 3 - Explosive Strength: This is the ability to do work in a burst of effort. This ability would be seen in the shuttle run test. 4 - Static Strength: This is the ability to exert force over a brief period. This ability would be seen in weightlifting. 5 - Dynamic Strength: This is the ability to exert force repeatedly or continually, as when doing pull-ups. 6 - Trunk Strength: This is the ability of the trunk muscles, especially the abdominal muscles, to exert force repeatedly or continually, as when doing leg-lifts. 7 - Gross Body Coordination: This is the ability to coordinate the activity of different parts of the body. 8 - Gross Body Equilibrium: This is the ability to maintain balance without visual input, as when standing on one leg with your eyes shut. 9 - Stamina: This is the ability to maintain maximum effort over a period of time.

 

Psychomotor Ability Factors: 1 - Control Precision: This is the ability to fine control rapid muscle adjustments. 2 - Multilimb Coordination: This is the ability to coordinate a number of limbs simultaneously. 3 - Response Orientation: This is the ability to select the right movement from a repertoire of several possible movements at speed. 4 - Reaction Time: This is the ability to respond quickly to a stimulus. 5 - Speed of Arm Movement: This is the ability to execute a rapid ballistic arm movement. 6 - Rate Control: This is the ability to make continuous anticipatory motor adjustments in a tracking task. 7 - Manual Dexterity: This is the ability to perform hand-arm manipulations of fairly large objects at speed. 8 - Finger Dexterity: This is the ability to perform finger manipulations of fairly small objects at speed. 9 - Arm-Hand Steadiness: This is the ability to make precise arm-hand positioning movements (as opposed to powerful or rapid ones). 10 - Wrist-Finger Speed: This is the ability to perform wrist-finger manipulations at speed. 11 - Aiming: This is the ability to position probes and pointers accurately at speed, or to dot in very small circles and suchlike.

 

There is, of course, no memory involvement whatsoever with the physical proficiency factors. These - as their name suggests - are purely skeleto-muscular in derivation, and any improvement with practice is entirely due to physiological factors (muscle strength, joint mobility, glucose metabolism, etc). The psychomotor factors, on the other hand, are mediated by the central nervous system, and when they show improvement with practice it has to be due - in part at least - to memory-like processes resulting from that practice.

 

But the situation is highly complex. For one thing, Fleishman describes the physical factor of gross body coordination in remarkably similar terms to the psychomotor factor of multilimb coordination, and yet the former is scoring a skeleto-muscular ability whilst the latter is scoring some aspect of nervous system organisation and efficiency. For another, there is evidence that the psychomotor factors are in fact largely separated from the "higher order" mental functions such as awareness and will. It is not uncommon, for example, to find that brain-injured subjects can learn new motor skills without being able to remember any of the practice sessions! And last but not least, the relative contribution of Fleishman's 11 psychomotor factors actually shifts as performance improves with practice. Thus when first learning a particular skill, the reaction time, speed, and dexterity factors effectively play no part at all: like Craik in his simulator, you are too busy getting things right to worry about doing them quickly. However, by the time the skill is well established, these same factors are contributing about 25% of the variance (Fleishman and Hempel, 1955).

 

The most probable reason for all this confusion is that the intact brain works by linking up separate systems, known as modules, each of which can be separately trained, and each of which can be separately damaged. And you become "skilled" when each module knows its own job and can talk fluently to its neighbour. Unfortunately, psychology has not yet figured out exactly how many modules there are, nor what each of them is responsible for, nor (consequently) what information needs to pass between them.

 

 

4 - Definitions of SA

As far as we can establish, the concept of SA was born within the specialist and rather secretive world of military ergonomics and air accident investigation in the mid-1970s, as an attempt to explain the large number of individual variables known to affect the cognitive performance of military and civilian aircrew. However, it did not go public in the civilian literature until about ten years later [the earliest PsycINFO reference when we checked 26th February 2003 was Spiker, Rogers, and Cicinelli (1986)]. It is best introduced by looking at some of the attempts to define it .....

 

SA is the ability to "maintain the 'big picture' and think ahead" (Dennehy and Deighton, 1997, pI.284).

 

"SA is defined as the 'operational space' within which personal and environmental factors affect performance" (Dennehy and Deighton, 1997, pI.287).

 

"[SA is] the degree of accuracy by which one's perception of his current environment mirrors reality" (US Navy website).

 

"[SA is a] cognitive state or process associated with the assessment of multiple environmental cues in a dynamic situation" (Isaac, 1997, pI.185).

 

"[SA] can be thought of as an internalised mental model of the current state of the flight environment. This integrated picture forms the central organising feature from which all decision making and action takes place. A vast portion of the aircrew's job is involved in developing SA and keeping it up to date in a rapidly changing environment." (Endsley, 1999, p257; emphasis ours.)

 

"[SA is] the accessibility of a comprehensive and coherent situation representation which is continuously being updated in accordance with the results of recurrent situation assessments" (Sarter and Woods, 1991, p52).

 

But the definition we like best (because it is closest to James's conceptualisation of consciousness) is .....

 

"[SA is] the perception of the elements in the environment within a volume of time and space, the comprehension of their meaning, and the projection of their status in the near future" (Endsley, 1988, cited in Endsley, 1999, p258; emphasis ours). 

 

SA is thus substantially more than the simple perception of data. Instead, it is an index of how a multitude of separate elements interact, and of how that interaction then affects what each element is likely to do next; it is William James's object implications all over again. And why does all this matter? Well above all because .....

 

"..... problems with SA were found to be the leading causal factors in a review of military aviation mishaps [citation], and in a study of accidents among major aircarriers, 88% of those involving human error could be attributed to problems with situation awareness" (Endsley, 1999, p258; emphasis added).

 

 

5 - SA as a Multidimensional Construct

The fact that SA is not a single ability creates serious problems for human factors researchers, because it means they have no single straightforward measure of the phenomenon they are most interested in. Indeed, when Dennehy and Deighton (1997) reviewed the literature from the period 1979 to 1992, they identified no less than 28 different variables which had been used as measures of SA. Suspecting that these variables could profitably be clustered, Dennehy and Deighton set out to find the main SA subscales. After discounting some obvious synonyms, they reduced the list of factors to 22 variables, and they then arranged for 231 pilots and instructors to rate these 22 variables on a seven-point scale of how necessary each was to maintaining an effective level of SA. A principal components analysis was then carried out on the raw subjective ratings, and this revealed five main subscales (the authors use the descriptor "meta-categories"), as follows .....

 

SA Meta-Category 1 - Pilot Knowledge: The first principal component was a composite measure of knowledge, per se.

 

SA Meta-Category 2 - Anticipation and Understanding of Future Events: The second principal component was a composite measure of an ability cluster for anticipating and understanding future events.

 

SA Meta-Category 3 - Capacity to Manage Stress, Effort, and Commitment: The third principal component was a composite measure of an ability cluster for managing stress and maintaining effort and commitment.

 

SA Meta-Category 4 - Capacity to Perceive, Attend, Assess, and Assimilate Information: The fourth principal component was a composite measure of an ability cluster for obtaining and processing information from the world itself, both directly, and from the cockpit dials and gauges available.

 

SA Meta-Category 5 - Overall Awareness: The fifth principal component was a composite measure of general awareness, similar to the "g-factor" popular among some intelligence theorists. 

 

In a further attempt to understand this clustering, and to help defend against criticisms of the subjective ratings methodology, Dennehy and Deighton adopted the Endler-Mischel "Interactionist Theory" (Endler, 1973; Mischel, 1973). This theory holds that it is impossible "to understand the role of an operator in isolation from the context within which she/he operates" (Dennehy and Deighton, p285). The result is a "P-E Fit" model, that is to say, one which tries to "fit" the person into his/her environment. However, the overlap between the person and the environment can never be complete, meaning that four distinct theoretical domains need to be taken into account, as follows .....

 

"Capacities" - P-Factors, Not Interacting: These are matters of fact relating to the person and defining what they are objectively capable of. This might include their baseline memory span or reaction time, for example. Dennehy and Deighton deal with this cluster of factors under the title "objective person" factors.

 

"Abilities" - P-Factors, Interacting with E-Factors: These are the environment-facing aspects of the person, such as task awareness and the ability to anticipate system response in a given environment. Dennehy and Deighton deal with this cluster of factors under the title "subjective person" factors.

 

"Roles of Work" - E-Factors, Interacting with P-Factors: These are the person-facing aspects of the environment, such as the tasks demanded of a system operator, or the objectives of their mission. Dennehy and Deighton deal with this cluster of factors under the title "subjective environment" factors.

 

"Real World" - E-Factors, Not Interacting: These are matters of fact relating to the environment, such as the objective nature of the equipment, the nature of the weather, etc. Dennehy and Deighton deal with this cluster of factors under the title "objective environment" factors. 

 

Dennehy and Deighton then characterised SA as the "operational space" provided by the two interacting domains. The thrust of this argument is summarised diagrammatically in Figure 1.

 

Figure 1 - The Interactionist ("P-E Fit") Model of Situational Awareness: Here is Dennehy and Deighton's "interactionist framework". It shows two main categories of factors, namely personal (or "P") factors (left oval) and environmental (or "E") factors (right oval). These separate sets of factors have been aligned so that they only partly overlap, giving a central zone (deeper yellow) where the P- and E-factors "interact" (hence the tag "interactionist"), plus two non-interacting zones on the flanks where objective measurement is possible. Specimen factors for each sub-cluster are shown from left to right across the belly of the diagram. The operational space which makes up SA is the zone of overlap between the two ovals. The inner dotted ovals show areas where the subjective and objective elements further interrelate.

If this diagram fails to load automatically, it may be accessed separately at

http://www.smithsrisca.co.uk/PICSA-fig1.gif

PICSA-fig1.gif

 

Enhanced from a black-and-white original in Dennehy and Deighton (1997; Figure 1). This version Copyright © 2003, Derek J. Smith.

 

Endsley (1995, 1999) divides the known factors up in a different way, separating them into (a) perceiving, (b) understanding, and (c) prediction factors, and naming the resulting categories as follows .....

 

"Level 1 SA": This level of SA contains the factors which affect the efficiency of the perceptual process.

 

"Level 2 SA": This level of SA contains the factors which promote an accurate current understanding of the world.

 

"Level 3 SA": This level of SA contains the factors which allow for accurate projections of future states or events.

 

However, Endsley then points out that it is simultaneously possible to divide up a pilot's declarative knowledge base (that is to say, Dennehy and Deighton's first meta-category) according to the type of knowledge involved, as follows .....

 

Geographical SA: This knowledge domain stores facts about the physical world, for example, the heights of mountains, the nature of the terrain, the locations of alternative airfields, the direction of the prevailing wind, etc. 

 

Spatial/Temporal SA: This knowledge domain stores facts about the four-dimensional physical world (the three spatial dimensions, plus time), for example, aircraft position and speed, time into mission, etc.

 

System SA: This knowledge domain stores facts about system states, for example, engine, control, or instrument malfunctions, fuel status, etc.

 

Environmental SA: This knowledge domain stores facts about weather and visibility, together with such things as restricted (keep out) or prescribed (keep in) airspace.

 

Tactical SA: This knowledge domain stores facts about the identity and capabilities of all other units in the vicinity (and, if military, that will include their combat intention as well).

 

Endsley brought all these concepts together into what he presents as an all-embracing theoretical model of SA, as now reproduced in Figure 2.

 

Figure 2 - Endsley's (1995, 1999) Model of Situational Awareness: Here is Endsley's version of the interaction between person and environment. The green-highlighted box represents the broad spread of the pilot's higher cognitive functions, and incorporates all three levels of SA (light blue panel, centre), plus his/her behavioural selection and execution capacities (see the "decision making" and "performance of actions" boxes). This higher cognitive resource is seen as being influenced from two directions, namely (1) the person (see the "individual factors" caption, bottom left), and (2) the environment (see the "task/system factors" caption, top left). The yellow-highlighted box (bottom right) represents more basic cognitive resources, and this includes the three main categories of knowledge, namely (a) procedural knowledge (the "information processing mechanisms" box), (b) declarative knowledge (the "long term memory stores" box), and (c) skill (the "automaticity" box) [this cross-mapping is ours, and approximate]. Note how matters of complex understanding ("goals and objectives") and misunderstanding ("preconceptions" and "expectations") have been given a box of their own (bottom left). This is not without justification, given the number of real life incidents in which preconceptions and false expectations have remained fatally uncorrected. Note also the four red "DANGER ZONE!!" panels. These mark points where the system is particularly vulnerable to error, and will be referred back to by subsequent discussion. 

If this diagram fails to load automatically, it may be accessed separately at

http://www.smithsrisca.co.uk/PICSA-fig2.gif

 

 

Enhanced from a black-and-white original in Endsley (1999; Figure 11.1) After Endsley (1995). Danger zoning overlays ours. This version Copyright © 2003, Derek J. Smith.

 

 

6 - When SA Fails

Endsley (1995, 1999) believes that the factors which most predict and limit SA performance are attention and working memory (WM) capacity, two of the most important functional elements within the "Information Processing Mechanisms" box in Figure 2. Endsley regards the limits of attention and WM as being largely constitutional, set ultimately by the fundamental layout of our nervous systems, and thus at best slow to improve with training. This takes us to the first of the danger zones overlaid onto Figure 2, namely "OVERLOAD ERRORS", and this is how Endsley assesses the risk so far as attention is concerned .....

 

"Direct attention is needed for perceiving and processing the environment to form SA, for selecting actions and executing responses. In the complex and dynamic aviation environment, information overload, task complexity, and multiple tasks can quickly exceed the aircrew's limited attention capacity. Because the supply of attention is limited, more attention to some information may mean a loss of SA on other elements. The resulting lack of SA can result in poor decisions leading to human error." (Endsley, 1999, pp260-261; emphasis added.)

 

SA can also be adversely affected by such factors as poor communication, fatigue, stress, both overhigh and overlow workload, inappropriate expectation, and even the feeling of being pressured to "press on regardless". Endsley (1999) summarises a number of earlier studies to provide an epidemiological view of SA errors, by level and subtype. This is summarised diagrammatically in Figure 3.

 

Figure 3 - The Epidemiology of Disasters: Here is a breakdown of US air accident statistics from Jones and Endsley (1995), broken down by SA Level and specific cause. Failures at SA Level 1 constituted 80.2% of the total, and fell into the five subtypes shown. The worst single culprit (at 37.2%) is information which is available to the pilot or crew, but which is not attended to for some reason (eg. due to distraction or excess workload). Failures at SA Level 2 include using "the wrong mental model" to interpret perceptual data, and this factor thus includes the sort of expectancy errors seen in the USS Vincennes [details] and Kegworth disasters [details].

If this diagram fails to load automatically, it may be accessed separately at

http://www.smithsrisca.co.uk/PICSA-fig3.gif

PICSA-fig3.gif

 

 

Enhanced from a black-and-white original in Endsley (1999; Figure 11.2). After Jones and Endsley (1995). This version Copyright © 2003, Derek J. Smith.

 

However, Sarter and Woods (1991) warn that there is no such thing as the most important factor. This is their argument .....

 

"Attempts to define the critical contents or components of situation awareness in general suffer from the fact that, given the dynamic environment of the flight deck, the relevance of data and events depends on their context, [and] will therefore vary within and between flights" (p47).

 

Hitchcock (1999) adds that despite the clarity of the data presented in Figure 3 we should nevertheless interpret the phrase "pilot error" wisely, even where there has been a demonstrably incorrect control action. This is because most such errors are in fact "mistakes waiting to happen" (Hitchcock, 1999, p311); it is just that the pilot in question was simply the person in the hot seat when things finally went wrong. Mistakes waiting to happen enter the system via the "DESIGN AND MANAGEMENT ERRORS" and "SELECTION AND TRAINING ERRORS" danger zones shown in Figure 2. 

 

 

7 - SA and Workload

Jensen (1997) is typical of recent work in this area. She warns that there is no satisfactory single measure of workload, especially for cockpit or process control tasks, and, like Dennehy and Deighton, has been forced to rely on subjective ratings. Working at the Farnborough DERA (= Defence Evaluation and Research Agency), she has combined self-report data with frame-by-frame video analysis and instrumentation data, and finds that if input workload, central processing, and output workload are monitored separately, they are found to respond differently to different types of event. In a simulated helicopter-helicopter combat, she found that central processing surges when tactical decision making - the command element of command and control - is called for.

 

Another way for workload to diminish performance is for it to impair communication. Two of Huber's (1982/2003 online) laws of organisational communication are worth quoting on this point .....

 

"The probability that a message will be transmitted from a unit is inversely related to the workload of the unit" (p143; Proposition R2).

 

"The probability or extent of message modification [later defined as the "undesired reduction of message complexity"] is positively related to the extent of the sender's work overload" (p148; Proposition M6).

 

 

8 - SA and Situation Assessment

Prince and Salas (1997) have studied how SA is formed in the first place, and then subsequently kept up-to-date. They argue that if SA is a complex mental state, maintained by a complex of cognitive processes, then problems with the updating processes are likely to be just as critical as problems with the state itself. Prince and Salas call the process(es) of continuous updating "situation assessment", and characterise the relationship between awareness and assessment as follows:

 

"Within the context of normal flight, situation assessment must occur continuously to ensure situation awareness. In addition, [] when a decision event arises, situation assessment must serve the decision process." (Prince and Salas, 1997, p291).

 

In their subsequent discussion of the awareness events, they cite a paper by Adams, Tenney, and Pew (1995), which emphasised the role of top-down influences in directing exploration of the environment. The two things go hand in hand .....

 

"..... the crew must have a good understanding of their present situation so they can assess the stimulus event for relevance, urgency, procedural, and goal related implications" (Prince and Salas, 1997, p293).

 

And in their discussion of the central decision making, Prince and Salas draw on the work of Lipshitz (1993), Klein (1993), and Noble (1993). Lipshitz had compared various competing theories of "naturalistic decision making", and found that situation assessment was a major factor in most of them. He defined the process as a "sizing up and construction of a mental picture of the situation" (Lipshitz, 1993, p132). Klein had discussed "recognition primed decision making", where "experience allows the decision maker to identify a single good option for action" (p292), and where assessment of the situation is critical to deciding which action from among many counts as "good". And Noble had introduced the term "reference problems" to describe memories for problems previously solved, proposing that they were indexed within the memory network on three major dimensions, namely (a) objective, (b) action, and (c) environment. Thus, one might choose the objective of "escape" rather than "surrender", the action of "run to the left" rather than "run to the right", and the environment of "jungle" rather than "desert". 

 

 

9 - SA, Risk Space, and Time to Contact

Another insight into situation assessment comes from considering humankind's evolutionary past. We evolved in a world full of predators, where being surprised would often mean being eaten and where only the fittest survived to breed. As a species, we responded to this challenge by going for cognitive processing power and teamwork rather than sharper teeth and longer claws. The behaviours which helped you make it through the day included bonding and social altruism, and the cognitive capacities included predictive feedback systems, motor schemata, the powers of abstraction and symbolisation, learning from experience, effective problem solving, plus one which is closely related to what we now call SA, namely "time to contact (TC)" processing. Smith, Lewin, and Hancock (1997) summarise the problem thus: "The detection and avoidance of collisions in a congested space is a problem that all independently mobile agents must solve quickly and efficiently" (pI.229). They then suggest that the single most valid factor in predicting safe navigation is ultimately an organism's estimate of TC. They studied this variable in commercial airliner pilots, and found that conflicts (ie. potential collisions) were consistently detected with no less than three minutes to spare as they entered a subjective "risk space". This indicates that experienced pilots code all items entering their risk space on a TC basis. There is however little in the literature on what the upper limit to this sort of target tracking is. As to the peripheral mechanisms involved, the work on "optic flow" fields may be implicated (see, for example, Koenderink, 1986, and Van de Grind, 1988, if interested).

 

 

10 - SA and the Mental Model

The modern concept of the "mental model" has been very heavily influenced by the work of Philip Johnson-Laird, Stuart Professor of Psychology at Princeton University. Johnson-Laird wrote his first major paper on the subject in 1981 (Johnson-Laird, 1981), and the monograph "Mental Models" in 1983 (Johnson-Laird, 1983). Nevertheless, the term goes back at least to the 1940s, for it appears in Craik's wartime papers, thus .....

 

"[While] as a rule mere objects evoke, in themselves, little response, anything which in the light of our previous knowledge puzzles us or defeats our insight or suggests new possibilities may evoke a very definite response. We want then to find some scheme in which our various experiences combine to produce patterns in us [..... which .....] can become developed to the point where their relevance to one another becomes functional; in consciousness this is indicated by ideas 'striking' us []. We then have a mental model of a possible event in the external world ....." (Craik, undated pre-1945, in Sherwood, 1966, p72; emphasis added.)

 

Similarly, nearly half a century later Endsley (1999) used the term "internalised mental model" in his definition of SA (see Section 2), and in explaining the influence of memory on current control behaviour. This is how he states the relationship:

 

"During active decision making, a pilot's perceptions of the current state of the system may be matched to related schemata in memory that depict prototypical situations or states of the system model. These prototypical situations provide situation classification and understanding and a projection of what is likely to happen in the future (Level 3 SA)." (Endsley, 1999, p263.)

 

But are Johnson-Laird's mental models and SA's "internalised mental models" one and the same thing? Well it appears not. The problem is that there are several qualitatively different types of memory, all interacting to add content to consciousness. Johnson-Laird and his followers have largely studied verbal reasoning and the mental model they have modelled is different from the sort of three-dimensional spatial intelligence measured by SA researchers. Helge Helbing, of the Technische Universität, Berlin, has studied the latter process in air traffic controllers (Helbing, 1997). The basic problem, Helbing asserts, is that while we know a lot about the mental structures which are activated during text comprehension, "we know relatively little about the cognitive representation of spatial information" (pI.177). Helbing goes on to cite earlier work by Whitfield and Jackson (1982), who had argued that what is most important to controllers is the "picture" they form and maintain. Helbing summarised Whitfield and Jackson's approach in the diagram now reproduced in Figure 4.

 

Figure 4 - The Mental Picture: Here is a 2S-O-2R variant of the classical S-R diagram [reminder], in which one of the main functions of central processing is characterised as maintaining a "picture" of what is going on in the world. Note how this diagram is structurally very similar to the core layout of the "transcoding models" commonly used within modern psycholinguistics [details]. The central understanding of the world - the "picture" shown below - is referred to therein as the "Semantic System" [see, for example, the Ellis and Young (1988) transcoding diagram].

If this diagram fails to load automatically, it may be accessed separately at

http://www.smithsrisca.co.uk/PICSA-fig4.gif

PICSA-fig4.gif

 

 

Enhanced from a black-and-white original in Helbing (1997; Figure 1). After Whitfield and Jackson (1982). This version Copyright © 2003, Derek J. Smith.

 

The work of Dr Dieter Wallach, of the Universität des Saarlandes, Saarbrücken is also relevant here. Wallach (1997) argues that cognitive engineering - applications of cognitive psychology within the engineering sciences - frequently involves problems of complex problem solving, and that it is commonly assumed that active interaction with a system is necessary to develop the mental model of that system required by its controller if s/he is to control it. But again different memory types allow different levels of performance. Wallach carefully distinguished between knowing how something works (which is deep conceptual knowledge) and knowing how to work something (which is far shallower practical knowledge), and stressed that the latter is sufficient only until things start to go wrong. Wallach's research therefore investigated the effects of different types of training activity upon the development of control room skills. The test environment was a computer simulation of the control room of a coal-fired power station, and the subjects were 40 engineering students divided into two groups. the first group - a system control group - was allowed to explore, interact with, form and test hypotheses about the system. The second group - a monitoring group - was only allowed to observe a trained operator at work: that is to say, it was given no "hands-on" experience. A battery of tests was then used to assess system knowledge (declarative knowledge) and control performance (procedural knowledge). Results indicated a significant superiority of the monitoring group, in sharp contrast to the prevailing assumptions. There was also a strong correlation between knowledge and performance in the system control group, but none in the monitoring group

 

 

11 - Shared SA (SSA)

More problems are likely to arise whenever SA becomes a team responsibility. This is because teams can all too easily add to their members' cognitive workload rather than reducing it. This is because any one individual's mental model has to include other team member nodes in his/her definition of the situation. A co-pilot, for example, needs a mental model in which the pilot's mental model is represented.

 

ASIDE: This is remarkably similar concept to the "theory of mind" literature which has accumulated in the last quarter century. The co-pilot is modelling a model [ie. meta-modelling] and researchers are modelling that meta-model [ie. meta-meta-modelling!!].

 

Yet again, the idea is a good one but the science is not yet in place to develop it. Nofi (2000/2003 online) warns that "the concept of 'shared situational awareness' [.....] is elusive and ill-defined, and does not lend itself easily to traditional scientific evaluation" (p1). Nevertheless, he manages to distill a dozen or so competing definitions into the following .....

 

"..... it develops by a process of integrating the mission-essential overlapping portions of the situational awareness of individual team members - thus developing a group dynamic mental model" (ibid, pp1-2).

 

"Crew Resource Management" (CRM) [previously Cockpit Resource Management] was introduced in the early 1990s to help overcome the limitations of individual understanding by broadening the concept of command and control to bring all flightdeck and cabin crew (and, on occasions, the passengers themselves) into the loop. [See our Transportation Disasters - Aerospace Database, under the headings Sioux City and Kegworth, for these are the disasters which most directly prompted this initiative.] And if CRM is as good as it has been claimed to be, then it follows that there should be measurable improvements in SSA as a result. This issue is still under investigation, and there is little available data at present.

 

Nor can the definition of "the team" even be restricted to those on the aircraft. Hill (2003 online) explains how traditional flight training fails to explain how to tell the important clues from the unimportant, given that flying exposes pilots to constant activity centred around their radio interaction with Air Traffic Control (ATC). He points to a particular weakness in training procedures, to the effect that pilots are insufficiently aware of how controllers do their job. This leaves them unable to see behind the words emanating from their earpieces, and thus less able to spot dangers in advance. Hill supports this argument with material taken from the radio interactions immediately prior to the 1991 Los Angeles International (LAX) collision between a US Air 737 and a Skywest Metroliner [details]. They established that the Skywest crew failed to note ongoing radio exchanges between the tower and the 737, even though they clearly stated that the 737 was intending to land imminently on the same runway they were about to take off from. They had, moreover, asked to join the runway from an unusual taxiway, and had been warned to wait for other ground traffic to cross in front of them. The end result was that none of the three separate control teams (that is to say, the 737 crew on their final approach, the Metroliner preparing to take off, and the controllers) had a comprehensive enough picture of the real world, only their confused assumptions. The 737 struck the Metroliner a few moments after touching down, and 34 people lost their lives [picture]. Hill's conclusion is that you must take your time and deliberately go as far beyond the minimum standards set by others. And yes, if that means putting a lot of effort into being your brother's keeper all the time, then so be it, because that is how the world works, and it is a lot better than getting fried. 

 

 

12 - SA and Automation

Endsley (1999) warns that it is all too easy for SA to be reduced by task automation. This is because it takes the human "out of the loop" (p267); worse - it renders them less capable of recovering effective manual control should the automation fail. This factor was clearly implicated in the 1987 Detroit MD-80 air disaster [details]. It is not all bad news, however, because automation of cockpit information flow, especially in the direction of greater integration and more efficient man-machine information transfer, can actually increase SA (although again you have to design in recovery pathways should the automation fail. Isaac (1997) adds that you also have to design so as not to disrupt visual search (because the visual iconic system relies on a memory type which is very sensitive to interference from near-simultaneous inputs) and minimises unnecessary delay (because short term memory is sensitive to decay).

 

 

13 - SA Outside the World of Aerospace

UNDER CONSTRUCTION 

 

 

14 - Training Implications of the SA Concept

Sarter and Woods (1991) warned that SA cannot be sustained over time unless pilots and crew are sensitive to minor changes in the massive volume of data presented to them by their instruments. They pointed to the fact that pilot training regularly simulates such "instant-onset failures" as engine failures, but devotes far less effort to training at detecting gradual deterioration of one data index amongst many (the same slow needle-creep which regularly causes motorists to run out of fuel or boil dry). Sarter and Woods refer to this type of event as "subtle problems".

 

Other researchers have studied process control industries such as power generation, petrochemicals, and steel manufacture. For example, Kluwe (1997) looked at the effects of different training methods on acquiring an industrial control performance and contrasted the two basic philosophies of training, namely learning by being told and learning by solving problem. The commonly claimed advantages of the latter, he claimed, were often more assumed that empirically demonstrated. The test environment was a computer simulation of an asphalt mixing plant, and 42 subjects took part. Initially, all 42 subjects were given equal experience with the system's training manual. Then they were split into two groups. Group 1 was the learning by being told group, and group 2 was the learning by solving problems group. Both groups were then given different experience for 6 days, and then tested. Results showed that the levels of achieved system knowledge did not differ, nor did control performance "on a global level". However, at a finer level of analysis it emerged that Group 1 was inferior when (a) subsequently having to acquire additional control skills (i.e., they were not as good at extending their knowledge), and (b) being required to modify their control strategy (i.e., they were not as flexible). This has considerable impact upon any training regime intended for operators in changing control environments.

 

 

15 - References

See the Master References List

[Home]