Course Handout - Transportation Disasters - Aerospace
Copyright Notice: This material was written and published in Wales by Derek J. Smith (Chartered Engineer). It forms part of a multifile e-learning resource, and subject only to acknowledging Derek J. Smith's rights under international copyright law to be identified as author may be freely downloaded and printed off in single complete copies solely for the purposes of private study and/or review. Commercial exploitation rights are reserved. The remote hyperlinks have been selected for the academic appropriacy of their contents; they were free of offensive and litigious content when selected, and will be periodically checked to have remained so. Copyright © 2010, High Tower Consultants Limited.
First published online08:33 BST 2nd May 2001, Copyright Derek J. Smith (Chartered Engineer). This version [HT.1 - transfer of copyright] dated 18:00 14th January 2010
Transportation Disasters - Aerospace
Key to Abbreviations:
AAIB = Air Accident Investigation Branch (UK)
CRM = Cockpit Resource Management (alternatively Crew Resource Management)
CVR = cockpit voice recorder
ENFL = English not first language
FAA = Federal Aviation Authority (US)
FDR = flight data recorder (or "black box")
GPWS = ground proximity warning system, a loud cockpit audio warning that an aircraft is flying dangerously low
HSC = Health and Safety Commission
knot = speed in nautical miles per hour, the standard measure of speed for aviation and marine purposes. The difference between knots and miles per hour (mph) arises because a nautical mile is 6076 feet, whereas a land mile is only 5280 feet. To get knots from mph, multiply the former by 5280/6076 (ie. roughly 0.87), and to get mph from knots, multiply the former by 6076/5280 (ie. roughly 1.15).
NTSB = National Transportation Safety Board
Staines Air Disaster, 1972: In this incident on 18th June 1972, a BEA Trident I stalled shortly after take-off from Heathrow en route for Brussels, and crashed into a field, killing all 118 people on board. The subsequent investigation found from the FDR that the droops (leading edge wing flaps) had been raised too soon, and that the built in "stick-push" safety mechanism had been manually overridden.
The aircraft left the ground at T+44 (seconds after brake release), making 145 knots. The seating on the flight deck was Captain Key at P1 (captain, front left), Second Officer Keighley, a trainee, at P2 (co-pilot, front right), Second Officer Ticehurst at P3 (behind), and Captain Collins at P4 (a non-participating transit passenger in the jump-seat behind P1). P1 and P2 shared the tasks associated with actively flying the aircraft, while P3 monitored systems and completed any necessary paperwork. In normal circumstances, either P1 or P2 can be the "handling pilot", with transfers of control made at the discretion or for the convenience of the captain; when trainees are involved, they are also selectively made to widen the trainee's experience. There is autopsy evidence that P1 may have been having, or may just have had, or may just have been about to have a heart attack, and it is perhaps relevant that he was a senior pilot with a reputation as a stickler for discipline. He was accompanied on the fatal occasion by two younger and much less experienced crew members, and by a fourth pilot due to fly out of Belfast later that day. He had also had a shouting match with a junior colleague one and a half hours before flying, which P2 had witnessed. This is what happened next according to the control tower recordings, the aircraft's FDR, and official expert interpolation; our own comments and interpolations are either indented or set in square brackets:
T+63 (FDR): The autopilot was engaged in the pitch and roll channels.
T+67 (FDR) - A SERIOUS ERROR: The autopilot was engaged in the airspeed channel, holding as far as turbulence allowed at the then current speed of 170 knots (instead of the recommended 177 knots). There is some suggestion that P1 habitually engaged the autopilot slow and early as a matter of personal professional preference.
T+72 (FDR): The aircraft started a 20° banked turn to port [and stayed banked until T+115, when as part of the developing emergency the port wing dropped momentarily and then levelled out].
T+83 (TOWER): P1 radioed "climbing as cleared".
T+90 (Presumptive): P2 announced "ninety seconds", indicating that noise abatement procedures needed to be invoked because the aircraft was about to overfly the built-up areas of the London commuter belt. These procedures required it to raise its trailing edge flaps and throttle back to a prespecified level. The leading edge droops, however, should remain in position (ie. lowered, or "out"). The location of the flap and droop control levers can clearly be seen at points 2 and 1 (respectively) on the AAIB photograph of the Trident control panel.
[T+92 (Presumptive): P1 verbally authorised flaps up.]
T+93 (FDR): Flying at 168 knots, the flap control lever was moved from the 20 DEGREES to the ZERO DEGREES position.
This operation would normally be carried out by P2 with his left hand. The mechanical process then normally takes around 10 seconds to complete, and reduces both lift (so that the margin of safety between the actual airspeed and the safe flying speed decreases) and drag (so that the engines can hold a given speed with lower throttle settings).
T+95 (FDR): The throttle settings were eased to the prespecified level [the throttling back was not intended to slow the aircraft, merely to reduce the noise it was making]. The location of the throttle levers can clearly be seen at point 3 on the AAIB photograph of the Trident control panel.
This operation, too, would normally be carried out by P2 with his left hand, grasping and moving all three throttle levers simultaneously, and would have taken several seconds of close attention to get the settings just right. As these changes to lift, drag, and power took effect, the autopilot would simply reduce the rate of climb to compensate, which may explain why the tailplane angle channel on the FDR shows a reasonably linear progressive increase in climb rate between T+90 and T+116, at which point the autopilot was disconnected in the developing emergency.
T+98 (FDR): With the flaps by now about 50% retracted, speed had dropped to 163 knots [perhaps because P2 had over-reduced the throttle settings].
T+100 (TOWER/FDR): P1 radioed "passing fifteen hundred". Speed at this point had increased to 166 knots.
T+103 (TOWER/FDR): With the flaps by now fully retracted, speed had dropped off again, this time to 157 knots. The tower responded with clearance to continue climbing to 6000 feet.
T+108: P1 acknowledged this clearance verbally, and P2 and P3 separately and successfully recorded it in their flight logs (Stewart, 1986). [P1's acknowledgement is on the control tower tapes. P2's and P3's logs were found in the wreckage, and were presumably updated (but not necessarily simultaneously) at some point between T+108 and T+115, when the warning lights starting to flash.] P1's message was the last transmission from the aircraft. The official report uses the adjective "terse" to describe it, and, in some experts' opinions, it came after a suspiciously long delay, possibly indicating that his medical condition was actively deteriorating. [For our own part, it seems just as compellingly to indicate P1 in mid-explanation to P2, or else assisting him to set the throttles - see our closing comments re over-attentiveness.]
T+108 - T+110: Speed blipped up from 157 knots to 163 knots and then down again to 157 knots [perhaps due to turbulence]. Several reports suspect a stick shake operation (see below) may have taken place at this juncture.
T+114 - THE FIRST CRITICAL ERROR: With an airspeed of 162 knots and at altitude 1772 feet, the droop select lever was moved to "up", beginning the process of winding up the droops themselves, a process which normally takes around 8 seconds and raises the safe flying speed from 177 knots to 225 knots. This immediately started to reduce the available lift, and ought to have been counteracted by lowering the aircraft's nose, increasing power, and reselecting droops down. Movements of undercarriage, flaps, and/or droops are called "changes of configuration", and if not properly allowed for and managed may induce a "configuration stall".
One Popular Scenario: Stewart (1986) suggests that P1 responded to a stick shake at around T+110 by saying "up" (meaning the throttles), but that P2 misinterpreted this instruction and raised the droops instead. However, it usually takes considerable detective work to determine why an emergency occurred and why it was not recovered from, and the fundamental choice here, as in all transportation disasters, is between (a) human error or inadequacy, and (b) mechanical failure. We shall now consider two issues simultaneously, namely who moved the droop lever, and whether P1 was medically incapacitated or not when it was moved. For the first of these issues there are only three possible solutions, namely (1) that P1 moved the lever, (2) that P2 moved the lever, or (3) that the lever moved itself (which is not as silly as it sounds - see under Schofield Theory below). For the second issue there are only two possibilities, namely (1) that P1 was incapacitated, and (2) that P1 was not incapacitated. This gives us a total of six permutations, as follows, four of which we are going to discard for the reason stated, and two we are going to retain for detailed analysis:
T+115: With the droops by now about 12% retracted (safe speed about 183 knots; actual speed about 162 knots; altitude about 1778 feet), the console DROOP WARNING (flashing amber) light would have come on. This warns about droops down at too high a speed or up at too low a speed, although with one's mind firmly in <ascent mode> there would have been no doubt which message was intended.
The location of the DROOP WARNING light can clearly be seen at point 6 on the AAIB photograph of the Trident control panel. The warning systems had been designed in consultation with the Applied Psychology Unit at Cambridge University, and were based on a Central Warning System (CWS) with 22 separate displays. Each pilot had general purpose amber and red warning lights to draw their attention to the CWS. In practice, therefore, pilots would get their initial warning from the console in front of them, but would then have to look to the centre line to see what the warning was about. On this occasion, the CWS would have displayed the message DROOP OUT OF POSITION.
T+115 1/2: With the droops by now about 17% retracted (safe speed about 186 knots; actual speed about 162 knots; altitude about 1791 feet), the stick shaker stall warning system would have engaged.
The textbook action at this juncture is known as "flying a recovery". This consists of three sequential, but fundamentally separate, cognitive operations, as follows, each with its own particular problems should they go wrong:
1 - Diagnosis Phase: With his/her mind now in <ascent mode with possible stall emergency>, the handling pilot must make an emergency diagnosis of the reason for the warning. If a stall warning is genuine it will be because the speed is too low given the attitude and configuration of the aircraft, which means checking airspeed, pitch, roll, and configuration, and deciding what - if anything - has gone wrong. If the diagnostic information confirms the warning, then the cognitive mode should immediately be changed to <fly stall recovery>, and the whole process will have taken not much more than a second. However, if the diagnostics are in any way unclear or inconsistent, then the emergency instantly becomes a <stall crisis> in which the reliability of every rule, control, subsystem, and display is in question, and in which survival becomes the only consideration.
2 - Planning Phase: Having entered <fly stall recovery> mode, the handling pilot now has to plan what s/he is going to do about it, and in what order. With most flight emergencies, of course, such plans have been well rehearsed in advance and will require little or no conscious thought (like braking if a child runs into the road in front of you). We have dealt elsewhere (Smith, 1997) with the process of micro-preparation for the execution of skilled motor movements, so suffice to note that this takes of the order of 170-210 milliseconds, depending on how many specific behaviours are being planned.
3 - Recovery Phase: With an appropriate plan of action in mind, the handling pilot now starts to execute the sequence of motor behaviours to bring it to fruition. With the case in question, this would be to lower the nose of the aircraft, increase the power, neutralise any roll, and change the configuration - but not necessarily in that precise order.
Here is how these three decision making stages might have worked themselves through in the two scenarios previously mentioned:
SCENARIO A - P2 MOVED THE DROOP LEVER BY MISTAKE; P1 HANDLING
In this scenario, P1 remains the handling pilot, but will not have seen P2 move the droop lever (or else he would have responded, and probably quite forcefully, at T+114):
1 - Diagnosis Phase: As an experienced pilot, P1's gaze will move in quick and accurate succession from the console warning light to the CWS diagnostic message (ie. DROOP OUT OF POSITION), and then to the droop control lever itself. If said lever shows UP, then he need look no further: he may accept the stall warning as genuine and enter the standard recovery mode.
2 - Planning Phase: The planning phase on this occasion would probably be as follows:
DROOP LEVER DOWN WITH RIGHT HAND
NOSE DOWN SIMULTANEOUSLY WITH LEFT HAND
THROTTLES UP WITH RIGHT HAND
3 - Recovery Phase: P1 now has to execute a thoroughly rehearsed and automatic series of recovery movements. However, the execution would have been overtaken by events at T+116, namely the pre-empting stick push, so the NOSE DOWN element would actually be done for him.
SCENARIO B - THE DROOP LEVER MOVED ITSELF; P1 HANDLING
In this scenario, too, P1 remains the handling pilot, but the major cause - the retracting droop - is not known to any of the flight crew because the lever moved itself:
1 - Diagnosis Phase: Same as Scenario A, but with a possible additional glance at the droop angle indicator after finding the lever UP.
2 - Planning Phase: Same as Scenario A.
3 - Recovery Phase: Same as Scenario A.
T+116: With the droops by now about 25% retracted (safe speed about 189 knots; actual speed about 162 knots, altitude about 1794 feet), the stick pusher stall recovery system operated, pushing the control column forward for about a second and then releasing it. this would have been accompanied by the console STALL RECOVERY (amber) light and a CWS STALL RECOVERY message. This action forced the nose of the aircraft downwards, but it would also have automatically disengaged the autopilot, giving an AUTOPILOT DISCONNECT (red flashing, with clang! clang! headset audio) warning to both P1 and P2.
There is then a three-second period in which the FDR shows the nose of the aircraft pitched - quite correctly - downwards. Coming off the top of its climb, it reached its maximum height of about 1800 feet at about T+117.
T+117 - THE SECOND CRITICAL ERROR: The control column was pulled back again, even though this meant fighting a built-in stall protection mechanism. As a result, by T+120 the aircraft had levelled out again. As a result, there was a second stick push at T+124, and a third at T+126, and on both these occasions the control column was pulled back afterwards. Finally, having inexplicably fought the stick push system three times in succession, the stall recovery system was manually inhibited at T+128.
T+130: The speed at this juncture was 193 knots, the highest value achieved on the entire flight, but well below safe flying speed. With the control column back, speed now fell away precipitously and the aircraft pitched nose-up, irrecoverably out of control.
T+150: The aircraft hit the ground, having been in the air less than two minutes.
If we examine this timeline, we see that the critical decisions were all concentrated in the 13 seconds between T+114 and T+127, and that the chances of successful recovery began to erode very rapidly after about T+120. This means that the crew had a mere 6 seconds in which to respond correctly or die, and had to spend the first two or three of those priceless seconds frantically checking instruments and controls. The incident is therefore an example of just how confusing it must be to have a sudden flurry of warning messages from both colleagues and systems, and to have only a second or two to diagnose the problem and plan a suitable remedy.
[In printing off a draft of this material on 24th April 2001 my printer flashed its OUT OF PAPER light. I turned and checked the INK level! Why? Because I always keep the paper well topped up, but use a lot of ink. It took me about 20 seconds to locate my error even though I had only two lights to choose from, and am a skilled printer operator! Had I been on Papa India that would have been enough time to kill me more than three times over!]
Here is the official list of five immediate causes of the crash:
(1) Captain Key failed to achieve and maintain adequate speed [ie. there was little or no margin for error - defence in depth]; (2) the droops were retracted some 60 knots too soon; (3) the crew failed to monitor the speed errors and to observe the movement of the droop lever; (4) the crew failed to diagnose the reason for the stick-push warning; (5) the crew cancelled ("dumped") the stall recovery system
And here is the official list of seven underlying causes of the crash:
(1) Captain Key's underlying heart condition lead to lack of concentration and impaired judgement; (2) Some distraction of P3 by the presence of the transit captain on the flight deck; (3) lack of training directed at "subtle" pilot incapacitation, that is to say, subacute and not generally apparent considerations such as that which might have been troubling Captain Key; (4) lack of experience in P2; (5) lack of knowledge in the crew relating to a "change of configuration stall"; (6) lack of knowledge that a stick-push would follow such an imminent stall; (7) lack of any mechanism to prevent retraction of the droops at too low a speed.
And here are some additional lines of enquiry pursued during the investigation and material to the recommendations made as a result of it:
"On experiment I find that if the droop lever is moved to the down position and the flap lever is moved before the droop lever is fully down, it is possible for the droop lever to become locked on the baulk only and not on the normal lock down. All appears normal  and it is not until the flap lever is selected up that anything untoward occurs. At this time the baulk is removed and the droop lever will return to the up position with the flap lever, the airspeed being some 50 knots below the correct speed for droop retraction." (AAIB, 1974, para. B.(vi).b2.)
In the event, however, the AAIB was satisfied this this could not have happened and in fairness it seems to make little difference to the success or otherwise of the emergency decision making as analysed in Scenarios A and B above.
RESEARCH ISSUE: There is little in the psychological literature on the normal recovery curve for extreme emotional states, nor how this might vary with personality. Captain Key, for example, had apologised to the colleague he had shouted at, but we do not know what his residual thoughts and feelings were. Nor were they necessarily negative, because it is recognised that some anger states result in a great deal of post-orgasmic guilt (eg. Warneka, 1998), so it could be that P1 was in fact being over-attentive towards P2 and deluging him with overly helpful explanations and demonstrations, rather than harumphing him. Which would explain the radio response delay at T+108.
The accident is significant because it forced the introduction of CVRs amongst UK airline operators, and is one of a series of accidents in which poor communication between captain and crew - specifically between an experienced P1 and a nervous P2 - was possibly a major factor. For further general details click here, and for the official AAIB report click here.
Air Accidents Investigation Branch (1973). Aircraft Accident Report No. 4/73. London: HMSO. [Available online.]
Smith, D.J. (1997). Human Information Processing. Cardiff: UWIC. [ISBN: 1900666081]
Stewart, S. (1986). Air Disasters. London: Arrow.
Warneka, T.H. (1998). The experience of anger: A phenomenological study. Dissertation Abstracts International: Section B: The Sciences and Engineering, 58(12-B June 1998):6863.
Paris DC10 Air Disaster, 1974: In this incident on 3rd March 1974, a Turkish Airlines DC10, crashed shortly after take-off killing 345 people. The immediate cause of the accident was the failure of a baggage handler to seal the cargo door properly. However, the fact that this sort of error should have been allowed to go undetected in the first place, and the relative severity of the resulting damage, was due to a string of earlier bad design decisions (Dixon, 1994).
Tenerife Air Disaster, 1977: In this incident on 27th March 1977, a KLM Royal Dutch Airlines 747 was taking off in fog from Los Rodeos Airport, Tenerife, having forgotten that a Pan-Am 747 was taxiing towards it on the same runway. The Pan-Am started to pull off the runway at the last moment, saving 77 of its passengers and crew, but the resulting near head-on collision nevertheless has the dubious honour of being the world's worst ever air disaster, with 583 killed. The subsequent NTSB, Spanish, and Dutch investigations failed to agree whether the accident was down to pilot error on the part of the KLM crew, or mis-instruction on the part of the control tower. There was certainly some suspicion that the KLM captain was eager to get away, and a deal of confusion in the interaction between both planes and the tower [for a transcript of the CVR, click here]. There was also some hint that the junior members of the flight crew were less than convinced that it was clear to take off, but only felt confident enough to hint at their reservations rather than stating them clearly. [For further general details, click here.]
Flight Palm 90, 1982: In this incident on 13th January 1982, an Air Florida 737 was unable to develop enough lift during a snow-bound take-off from Washington National Airport, and crashed into the ice-covered River Potomac. Five survivors were rescued from the river, but the remaining 75 passengers were killed. The subsequent NTSB investigation noted that the engines had not been set to full power because of false readings from iced up sensors. They also noted a deal of nervous banter on the CVR prior to take-off, and in their final judgement noted that the de-icing equipment had not been switched on. This incident is significant because it is one of a series including Staines (1972) and Tenerife (1977) where communication on the flight deck, especially from co-pilot to captain was less than totally effective. There was a distinct hint that the second in command on this occasion was less than happy with the decision to go ahead. [For further details and an audio extract from the CVR click here, and for a partial CVR transcript click here.]
Osaka Flight 123 Air Disaster, 1985: In this incident on 12th August 1985, a Japan Airlines 747 en route from Tokyo to Osaka lost its tail fin and hydraulic systems after its rear pressure bulkhead blew out. Out of control, it crashed into the Mount Osutaka, north west of Tokyo, killing all but 4 of the 524 people on board (making it the world's largest single aircraft air disaster). The cause of the accident was traced to a faulty repair to said bulkhead in 1978 after an earlier heavy landing. The repair design was sound, but its execution was defective (a single row of rivets was used instead of the three rows called for in the repair specification). However, the incident is not just a good example of the need for thorough quality control procedures during the engineering process, but also of how not to manage a disaster once it has happened. Here are three of the critical issues:
[For further details, click here.]
Challenger Space Shuttle Disaster, 1986: In this disaster on 28th January 1986, NASA Flight 51-L - the Challenger space shuttle - was destroyed in a total loss explosion 73 seconds into its mission. The immediate cause of the explosion was a failure of the O-ring seals between the lower two segments of its starboard solid rocket booster (SRB). All seven members of the crew were killed. A by-the-millisecond account of the flight is available on the NASA website.
It emerged early in the accident investigations that there had been a major problem with the NASA decision making process prior to the launch. The SRBs were manufactured by Morton Thiokol Inc., Utah, (henceforth referred to as Thiokol), and it was their long-standing and clearly expressed engineering opinion that the specification for the SRB rendered them unsafe for use at temperatures below 53º Fahrenheit. The O-ring segment-to-segment seals simply lacked the necessary resiliency to seal safely at lower temperatures than this. Yet the predicted temperature for the morning of the launch was 25-29º. [In the event, the ambient air temperature was 36º, and the temperature on the shady side of the SRB (the point where the fatal rupture occurred) was estimated at 28º plus or minus 5º.] But the launch had nevertheless been authorised.
The enquiry report concluded:
"1. The Commission concluded that there was a serious flaw in the decision making process leading up to the launch of flight 51-L. A well structured and managed system emphasising safety would have flagged the rising doubts about the Solid Rocket Booster joint seal. Had these matters been clearly stated and emphasised in the flight readiness process in terms reflecting the views of most of the Thiokol engineers and at least some of the Marshall engineers, it seems likely that the launch of 51-L might not have occurred when it did.
2. There was no system which made it imperative that launch constraints and waivers of launch constraints be considered by all levels of management.
3. The Commission is troubled by what appears to be a propensity of management at Marshall to contain potentially serious problems and to attempt to resolve them internally rather than communicate them forward. This tendency is altogether at odds with the need for Marshall to function as part of a system working toward successful flight missions, interfacing and communicating with the other parts of the system that work to the same end.
4. The Commission concluded that the Thiokol management reversed its position  at the urging of Marshall and contrary to the views of its engineers in order to accommodate a major customer."
[For the fuller text, click here, and follow the menu.]
And here are the main lessons of this accident:
For further general commentary, click here.
Habsheim A320 Air Disaster, 1988: In this accident on 26th June 1988, an A320 Airbus attempted a low and slow display flight across the Habsheim airshow in June 1988, but was unable at the end of the circuit to gain sufficient speed to clear a line of trees across the end of the runway. This was because the pilot - knowing that the A320 had a highly computerised flight deck - had purposefully disabled the GPWS. "He was so used to the system keeping the aircraft safe," says Race (1990), "that he felt it could wave a magic wand and get him out of any problem." It could not. The pilot's call for full power came too late and the engines were still accelerating when the aircraft hit the trees. Race's conclusion is ominous - complex systems routinely encourage overdependence on the part of their users. "The better our systems protect our clients," he writes, "the more likely it is that when a situation occurs outside the system's competence  the client will make a hash of things" (p15). This incident provides a good example of both a mode error, and of the problems encountered in producing safety critical software. [For an introduction to the cognitive science of mode error, click here.]
Race, J. (1990). Computer-encouraged pilot error. Computer Bulletin, August 1990, 13-15.
USS Vincennes Air Disaster, 1988: Because this is essentially a military command and control disaster, it is dealt with in the section on military systems failures. [To be transferred, click here.]
Kegworth Air Disaster, 1989: In this incident on 8th January 1989, a British Midland 737 en route from Heathrow to Belfast suffered a turbine failure in its port engine and was diverted for an emergency landing at East Midlands Airport, near Nottingham. What happened next is an excellent example of how cognitive failures can crash an aircraft.
The aircraft took off at 1952hr, and climbed normally for 13 minutes before the engine failed. This event was accompanied by smoke and vibration, and was wrongly judged to have taken place in the starboard engine. The AAIB report blamed this misjudgement on a combination of vibration, noise, and smoke outside the flight crew's training and experience. The critical factor here was that on older 737s the air conditioning intakes had been situated on the starboard side. However, the air conditioning system had been redesigned on newer aircraft. It now included a port side air intake, but that fact had gone largely unreported. When the crew smelled smoke, therefore, their immediate - but false - presumption was that it must be the starboard engine which had the problem.
Both engines were then throttled back and the worst of the smoke and vibration died away. In fact, this was coincidental - it was just that the damaged port engine chose that time to stabilise itself temporarily. Yet it helped to convince the crew that their initial diagnosis had been correct. The only inconsistency was that the port engine was showing vibration on the cockpit vibration gauges, whereas the starboard was not. However, that data was disregarded as unreliable, because the vibration gauges on older 737s had acquired a "common knowledge" reputation for being unreliable. Flight crew had grown accustomed to not relying on them. And again, when the gauges had been upgraded on newer aircraft, and now told the truth, that fact had gone largely unreported. The good starboard engine was then shut down altogether, and the port engine continued to turn, but under a slightly reduced load since the aircraft was now descending.
The flight crew's attention was now entirely taken up with weather reports, approach instructions, and carrying out the checklist for emergency landings with one engine. They also took time to make an announcement over the cabin address system to explain what had happened - that the starboard engine had been shut down, and that all was in hand to make an emergency landing. This announcement greatly puzzled and concerned the many passengers who had seen smoke and debris coming from the port engine, but they said nothing.
And then, when making the final approach, the stricken port engine failed completely. At this point, the aircraft was still two and a half miles from touchdown, and had already descended to 900 feet. There was accordingly insufficient time to restart the starboard engine, and 59 seconds later the unpowered aircraft hit the ground just short of the M1 motorway, bounced across it, and came to a halt on the embankment on the far side. Out of 126 passengers and crew on board, 47, all passengers, lost their lives.
The AAIB report also listed:
As a result of this incident, pilot training has been extended to include major specification enhancements and defect corrections, so that "known defects" are explicitly removed, and cabin crew are required to be consulted for their perspective on events. [For further details, click here .] [For an introduction to the cognitive science of situational awareness, click here.]
Sioux City Flight 232 Air Disaster, 1989: In this incident on 19th July 1989, a United Airlines DC10 en route from Denver to Chicago suffered a turbine failure in its tail engine, as a result of which all three sets of hydraulic control lines to the rudder and elevators were severed. The plane was then essentially unsteerable: all it could do was go faster and slower (and thus up and down). However, the residual throttle control did allow power to be reapportioned between the two wing engines, thus giving a very rudimentary ability to power steer, and this allowed the flight to divert for an emergency landing at Sioux City, Iowa. The crew were also lucky to have an experienced pilot amongst the passengers, who volunteered to take over the throttle steering while the regular crew went about the more urgent business of arranging for the crash landing. Nevertheless they were faced with solving a problem for which there existed no laid down procedures, because neither Boeing nor the airline had predicted loss of all three hydraulic systems.
The combined crew of four had no less than 103 years flying experience between them, and used it to the very best effect! They also benefited from a United Airlines training programme called Cockpit Resource Management (CRM) introduced in 1980 to improve effective emergency problem solving. By pooling their experience the crew managed to stabilise the aircraft, descend, and line it up with one of the runways at Sioux City. Even so, the final touchdown - without the benefit of flaps - was at 215 knots instead of the normal 140 knots, and in the resulting crash landing, 107 passengers were killed. The flight crew were among the 189 who survived. [For further details, click here and here. Cockpit Resource Management may also be referred to as Aircrew Coordination Training (ACT) or Crew Resource Management.]
Long Island Flight 052 Air Disaster, 1990: In this incident on 25th January 1990, a Columbian Avianca Airlines 707 ran out of fuel on its final approach to Kennedy International Airport, New York, and crashed killing 73 of its 158 passengers. Investigations revealed that the aircraft had informed ATC that it was low on fuel, but had failed to declare a formal emergency when its landing was further delayed by bad weather. Analysis of the CVR indicates a lot of discussion amongst the flight crew on whether or not ATC fully appreciated the seriousness of their plight, but nevertheless communication with ATC was only occasional and disturbingly unassertive. The FAA subsequently suggested greater standardisation of vocabulary and phraseology between flight crew and controllers to minimise the likelihood of further misunderstandings of this sort, especially where the flight crew were ENFL. [For general details click here, and for the CVR transcript click here.]
Bangalore Air Disaster, 1990: In this incident on 14th February 1990, an India Airlines A320 crashed on final approach to Bangalore airport, having undershot the runway by half a mile, despite perfect visibility. All 90 persons on board were killed. The probable cause of the disaster was the fact that the automated control systems had been accidentally set to <OPEN DESCENT> mode, which cuts the engines to idle and then automatically loses height to maintain speed. The crew were unable to diagnose their error quickly enough to take corrective action. This incident is now regarded as one of the classic examples of "mode error", one of the main problems encountered when automating systems control. [For an introduction to the cognitive science of mode error, click here.]
Los Angeles Air Disaster, 1991: In this incident during the night of 1st February 1991, an air traffic controller at Los Angeles International Airport (LAX) cleared a US Air 737 to land, having forgotten that a few minutes previously she had cleared a Skywest Metroliner to take off from the same runway. Due to contributory lack of radio monitoring on the part of the two flight crews involved, this error was not detected, and moments after the 737 touched down it collided with the stationary Metroliner, killing all 12 people on board. Severely damaged, the 737 then careered off the runway and hit a building, killing 22 of its own passengers [picture]. The NTSB report blamed "the failure of the Los Angeles Air Traffic Facility Management to implement procedures that provided redundancy  and the failure of the FAA Air Traffic Service to provide adequate policy direction and oversight to its air traffic control facility managers. These failures created an environment in the Los Angeles Air Traffic Control tower that ultimately led to the failure of the local controller 2 (LC2) to maintain an awareness of the traffic situation, culminating in the inappropriate clearances . Contributing to the cause of the accident was the failure of the FAA to provide effective quality assurance of the ATC system." (NTSB/AAR-91/08.) [For an introduction to the cognitive science of situational awareness, click here.]
Dahran Scud Attack, 1991: In this incident on 25th February 1991, a Raytheon MIM-104 surface-to-air missile (better known as a "Patriot") [pictures] was fired from Dahran, Saudi Arabia, to defend against an incoming Scud missile. The interception failed, and, by a fluke of targeting, the Scud hit a US barracks, killing 28 servicemen [details]. Failures such as this constitute operationally highly sensitive information, and so several deliberately conflicting cover stories were immediately released [example], but by 1992 investigations had revealed that the real cause of the incident was a "rounding error" in the missile's floating point arithmetic software (Skeel, 1992/2003 online). The US government's General Accounting Office reported that "the Patriot's weapons control computer used in Operation Desert Storm is based on a 1970s design with relatively limited capability to perform high precision calculations [.....] the conversion of time from an integer to a real number cannot be any more precise that 24 bits" (GAO Report B-247094, 1992/2003 online). [For more on the technicalities of rounding error in floating point arithmetic, click here and go to Section 1.3.]
Strasbourg A320 Air Disaster, 1992: In this incident on 20th January 1992, an Air Inter A320 crashed on a night-time approach to Strasbourg airport, after the crew had failed to detect they were descending too rapidly. As with the Bangalore disaster two years previously (above), it is possible that the inherent risks of the A320's <DESCENT> mode had not been fully appreciated. [For an introduction to the cognitive science of mode error, click here.]
Nagoya A320 Air Disaster, 1994: In this incident on 26th April 1994, a China Airlines A300 crashed on approach to Nagoya airport, following a clash of wills and understanding between the human crew and the autopilot. The autopilot won. All but 15 of the 279 people on board died. [Report] [For an introduction to the cognitive science of mode error, click here.]
Flight ZD576 Helicopter Air Disaster, 1994: This was a total loss of a military Chinook helicopter on 2nd June 1994. The aircraft crashed into a hillside on the Mull of Kintire, Scotland, en route from Northern Ireland. The 4 crew and 25 passengers (all police and intelligence service personnel) were killed. There has been considerable debate over the cause of this incident with concern over the safety of the new engine control software, FADEC. This system controls the flow of fuel to the aircraft's two engines, but is so complicated that it is virtually impossible to simulate all possible operational conditions. Along with the Habsheim (1988), Strasbourg (1992), Nagoya (1994), Fox River Grove (1995), and Ariane (1996) disasters, this incident is another good example of the problems engineering safety critical software.
Ariane 5 Flight 501, 1996: This was a total loss rocket launch on 4th June 1996. The vehicle veered out of control 36.7 seconds after lift-off, due to a failure of the back-up and main guidance systems. This put the steering hard over in an attempt to cure a defect which did not in fact exist. This, in turn, overstressed the airframe and caused the booster rockets to start to break away, and the resulting self-destruct explosion destroyed the vehicle. The guidance failure was caused by a software incompatibility: programs written for the Ariane 4 rocket series contained code not needed by, and catastrophically not compatible with, the Ariane 5 series.
Paris Concorde Air Disaster, 2000: In this incident on 25th July 2000, an Air France Concorde ruptured a fuel tank during take-off and crashed in flames a few seconds. See the referenced website for the principal facts and figures.
THIS SPACE RESERVED. BUT NOT FOR YOU, WE HOPE .....