Human Factors ATSB
Human Factors ATSB
This report was produced by the Bureau of Air Safety Investigation (BASI), PO Box 967, Civic Square ACT 2608.
Readers are advised that the Bureau investigates for the sole purpose of enhancing aviation safety. Consequently,
Bureau reports are confined to matters of safety significance and may be misleading if used for any other purpose.
As BASI believes that safety information is of greatest value if it is passed on for the use of others, readers are
encouraged to copy or reprint for further distribution, acknowledging BASI as the source.
Contents
Page
Abbreviations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv
Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . v
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
Aims . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
Analysis of incidents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
Safety actions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
Appendixes
1. Maintenance incidents. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
2. Human factors in maintenance coding sheet. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
3. Reliability of coding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
4. Definitions. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
5. Maintenance incident events . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 40
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
iii
Abbreviations
AME Aircraft Maintenance Engineer
ATA Air Transport Association
CAA Civil Aviation Authority
CFIT Controlled Flight Into Terrain
CRM Crew Resource Management
ETOPS Extended Range Twin-engine Operations
FAA Federal Aviation Administration
ICAO International Civil Aviation Organisation
LAME Licensed Aircraft Maintenance Engineer
MEDA Maintenance Error Decision Aid
OH&S Occupational Health and Safety
SHEL Software Hardware Environment Liveware
TCAS Traffic Alert and Collision Avoidance System
iv
Summary
Maintenance incidents contribute to a significant proportion of worldwide commercial jet
accidents, yet until recently, little was known of the nature of maintenance incidents and the
factors which promote them.
In face-to-face interviews, maintenance technicians were asked to report examples of maintenance
incidents which they had experienced first-hand. Eighty-six incident reports were recorded.
Human factors were involved in most of the reported incidents, with workers on duty between
the hours of 0200 and 0400 having a greater chance of having an incident than workers on duty
at other times of the 24-hour clock. The frequency of incidents increased as the shift progressed
up to the second-last hour, after which the frequency of incidents diminished.
For those incidents which had the potential to affect the airworthiness of an aircraft, difficulties
with procedures emerged as the most significant factor. This included misunderstandings and
ignorance of procedures.
For those incidents which had the potential to affect the health and safety of workers, difficulties
with tools and equipment emerged as the most frequent factor.
The majority of the human errors involved in incidents were rule-based mistakes, many relat-
ed to mistaken assumptions. Absent-minded slips and lapses were involved in approximately
one-third of the incidents.
The final section of the report contains suggested safety actions, intended firstly to reduce the
frequency of human error and maintenance incidents and secondly, to reduce the conse-
quences of any such errors which do occur.
v
Introduction
Figure 1 Safety issues versus onboard fatalities: worldwide jet fleet 1982–1991
Safety issue
Controlled flight into terrain (CFIT) 2169
Maintenance and inspection 1481
Loss of control 1387
Air Traffic Control 1000
Approach and landing without CFIT 910
Post-crash smoke or fire 644
In-flight smoke/fire 610
Ground de-icing/anti-icing 384
Windshear 381
Uncontained engine failure 199
Out of configuration takeoff 188
Airport ground operations 136
0 500 1000 1500 2000 2500
Onboard fatalities
Source: Boeing
Maintenance incidents are not merely costly in terms of life and property, but can also impose
significant costs when flights are delayed or cancelled. In 1989 maintenance constituted 11.8%
of US airline operating costs or more than US$8 billion per year (Shepherd 1992). The annual
cost to the Australian airline industry is likely to be in the order of several hundred million dol-
lars per year. Gregory (cited in Marx & Graeber 1994), estimated that each delayed aircraft costs
an airline on average US$10,000 per hour, while each flight cancellation can be expected to cost
approximately US$50,000. When these costs are considered, it is apparent that airlines stand to
gain significant benefits by even a small reduction in the frequency of maintenance-induced
delays, particularly those which occur closest to scheduled departure times, during line main-
tenance or when an aircraft is being prepared for departure.
Human error
Most accidents to complex industrial systems such as powerplants or transport systems feature
some involvement of human error. The terms ‘error’ and ‘human error’ are widely used in the
safety field and do not imply that operators are blamed for workplace incidents. Nevertheless,
concluding that human error was involved in an accident or incident does not generally help
to prevent such occurrences from happening again. In order to more fully understand why the
event occurred, it is necessary to describe in some detail the type of error the person made, and
if possible, to identify some of the reasons or factors which led to the error. Different types of
errors may require different preventative strategies.
One of the most basic description systems for human error is to categorise errors as those of
omission, commission or substitution, based on the work of Swain (Miller & Swain 1987). An
error of omission occurs when a person fails to perform a step in a task which should have been
performed; an error of commission occurs when a person performs an action which should not
have been performed. A related type of error is substitution, where an undesired action is per-
formed in place of the desired action. A fourth category of mis-timed actions can also be
included. This simple system of describing errors has the advantage of being relatively straight-
forward, but unfortunately does not describe the error in detail and does not give much insight
into why the person made the error.
A more detailed system of describing errors which gives insight into why the person made the
error was developed by Rasmussen, partly as a result of his examination of the errors of nuclear
powerplant operators. Rasmussen proposed that performance can be categorised according to
the level of cognitive control which the person is expending on the task at the time of the error.
2
His skill-rule-knowledge framework (1983) has become a widely accepted model of human
error. Using this framework, the activities of a maintenance engineer can be divided into three
types of actions: knowledge-based behaviour, rule-based behaviour and skill-based behaviour.
1. Skill-based behaviour
Skill-based behaviour is unconscious, rapid, does not seem to take conscious mental effort and
most importantly, is automatic. Many skilled routine actions such as opening and closing pan-
els can be performed automatically. Skilled workers possess an extensive repertoire of skill rou-
tines which can be initiated consciously and then left to run their course. Any maintenance task
which is performed frequently is likely to involve skill routines. One of the most common skill
errors is ‘environmental capture’ or habit intrusion. This occurs when a well learnt routine
action is performed in familiar surroundings, despite an original intention to perform anoth-
er action. A person who is distracted may carry out a well learnt action without modifying it
to new or unusual circumstances. A frequent error of this type is filling in a cheque in January
and writing in the previous year.
Another common skill error is the ‘omission following an interruption’. If a well-practised rou-
tine is interrupted, it may never be completed, or may be picked up again at the wrong stage.
One of the most dangerous manifestations of this in aviation is the interrupted checklist.
Omissions following interruption have particular relevance to aircraft maintenance.
Skill-based performance calls on very little mental effort and generally results in few errors. The
automatic nature of skill frees workers to think about other things, but the cost of this is that
they are less likely to monitor what they are doing. As a result, absent minded slips and lapses
are a particular risk. It is very difficult to modify an automatic skill once it has been learnt.
However, this is not to say that checks cannot be built into work performance.
A further difficulty with skill-based performance is that skilled operators are generally unaware
of the automatic procedures they are following and may be unable to explain to another per-
son how the task is performed.
2. Rule-based behaviour
People use rules or plans constantly in everyday life, without necessarily being aware of them.
These rules are often procedures which have been learnt through trial and error and are then
applied to situations as an aid to decision making in an ‘if...then...’ manner, for example, ‘If the
dipstick indicates that the engine oil is low, then top up the oil’. Although it is often a conscious
process, it does not require the person to go back to first principles in the way that knowledge-
based behaviour does. Aircraft mechanics constantly apply rules or expertise which enable
them to deal effectively with familiar or common situations. Many of these rules are formally
laid-down procedures; however, just as important are the unwritten work practices which are
applied to particular situations. Rule-based errors may occur when a person applies an inade-
quate rule to a situation or misapplies a good rule.
3. Knowledge-based behaviour
Knowledge-based behaviour is required when there is no pre-packaged solution to a situation.
Knowledge-based behaviour tends to be slower than other forms of behaviour and is very
demanding of mental resources, but is necessary when a person is faced with an unfamiliar
problem. Knowledge-based errors are errors of decision making, and may reflect a lack of
information on the task.
The skill-rule-knowledge distinction helps to explain why errors occur and to predict the types
of errors that will occur under various circumstances. Errors at the skill-based level are com-
monly referred to as slips and lapses, while errors at the rule-based and knowledge-based levels
are commonly referred to as mistakes. Many of the same basic error types have been observed
in a wide range of industrial settings, supporting the view that human errors are not usually
random deviations from normal performance, but rather follow systematic patterns and hence
can be partly predicted and prevented.
3
The Reason model of accident causation
In addition to the immediate unsafe acts or errors committed by operators, investigations typ-
ically reveal that longstanding systemic failures have had a role in causing, permitting or exac-
erbating accidents and incidents. Therefore it is important to consider not just the immediate
circumstances of maintenance incidents, but also the underlying or systemic failures which
make such incidents possible.
The model of system breakdown proposed by James Reason, illustrated in Figure 2, has become
a standard framework for analysing accidents in industrial and transport settings (see Reason
1990, 1991). The model has been advocated for accident investigation purposes by ICAO
(International Civil Aviation Organisation 1993), has been used in the analysis of anaesthetic
accidents (Runciman and others 1993) and has been applied by its originator to the analysis of
accidents in various settings, including nuclear power plants, chemical plants and transport
applications . While the Reason framework was initially proposed to account for accidents, it
can also be applied to less catastrophic occurrences.
Defences
Reason recognises that human behaviour is the greatest contributor to system failure. He con-
siders that system breakdowns result from combinations of active failures and latent failures,
sometimes in conjunction with unusual environmental forces. Active failures are the events
which immediately precede the breakdown. Unsafe acts such as errors or violations are the
most commonly identified active failures.
In Reason’s terminology, latent failures are the longstanding system problems which create the
circumstances in which active failures occur, and have the potential to make the consequences
of active failures especially serious. Latent failures include inadequate defence systems and con-
ditions which promote unsafe acts in the workplace. Latent failures often have their origin in
management and may be put in place well before the breakdown occurs. Using a medical anal-
ogy, Reason has given the label ‘resident pathogens’ to longstanding system failures.
4
In contrast to the Reason model, the SHEL model is most useful in considering human factors
at the ‘sharp end’, that is, the performance of individual operators. The SHEL model and the
Reason model complement each other and can be used jointly.
The SHEL model enables human factors issues to be divided into four broad areas. The first is
the interaction between people (‘liveware’) and software such as procedures, documentation
and manuals. The second element of the model is the interaction between people and hard-
ware, such as tools and equipment. The third element of the model represents the interaction
between people and the environment. The last element of the model represents the interactions
between people in the system, and includes issues such as communication, teamwork and
group interactions.
The SHEL model provides a simple but powerful framework in which most individual human
factors problems can be described. The SHEL model provides a useful guide to assist in the
investigation of maintenance incidents as it acts as a prompt to ensure that all relevant factors
have been identified.
5
Aims
This study was conducted with the aim of achieving a better understanding of maintenance
incidents, including the role of human error and the underlying factors which lead to such inci-
dents; identifying areas where safety improvements can be made; and making recommenda-
tions to achieve improvements.
6
Method
Maintenance technicians were asked to provide examples of incidents, following a structured
questionnaire based on the critical-incidents technique pioneered by Fitts and Jones (1947) in
their study of pilot error and further developed by Flanagan (1954). The critical-incident tech-
nique is a key human factors method which has been used for many years in the aviation indus-
try, and has also been applied in medicine and the nuclear power industry. The technique
involves gathering first-hand accounts by operators of critical incidents, accidents, mistakes
and near accidents which have occurred in the performance of job tasks. The technique is par-
ticularly useful where a system has been in operation for some time and where operational dif-
ficulties have been experienced but where the nature of the difficulties is not well understood.
A copy of the questionnaire can be found at appendix 1.
During interviews held in 1994 and 1995, maintenance technicians were asked to report inci-
dents which had occurred in the previous 12 months in which they had a first-hand involve-
ment, either as a participant or as an observer. All participants in this study were involved in
the maintenance of aircraft with a certified maximum seating capacity greater than 38 seats
and/or a maximum payload exceeding 4,200 kg. Although the primary focus of BASI is on
incidents which could affect the safety of public transport operations, incidents which could
have affected the safety of workers were also collected. Both types of incidents can arise from
underlying deficiencies in maintenance organisations.
Safety incidents frequently involve a sequence of events, so the incidents were broken down
into event sequences using a system developed by Williamson and Feyer (1990) at the National
Institute of Occupational Health and Safety. The coding system used to analyse incidents can
be found at appendix 2. This system was originally developed to examine fatal workplace acci-
dents. As well as considering each incident as a sequence of events, each event was categorised
according to whether it involved the actions of a person, a failure of equipment or an environ-
mental event such as wind or rain. The contributory actions of people in the incident sequence
were categorised according to Swain’s omission, commission, substitution categories of human
error and the SRK framework of Rasmussen (1983). Each event was considered as a separate
‘sub-incident’ and was assigned contributing factors where appropriate. A sample of the inci-
dents was coded independently by two coders to evaluate the reliability of the coding system.
The results of this evaluation can be found at appendix 3. Definitions of terms used during cod-
ing can be found at appendix 4.
7
Results
General
Of the 86 incident reports collected, 46 were classified as airworthiness occurrences, as they had
the potential to affect the operation of an aircraft, and 49 were classified as occupational health
and safety (OH&S) occurrences as they related to the health and safety of maintenance per-
sonnel. Nine incidents fell into both categories, in that they related to both airworthiness and
worker safety. A significant number of incidents (54%) had not been reported through official
channels. Error types were categorised with a system developed from Boeing’s Maintenance
Error Decision Aid (MEDA) system (Boeing 1994). A brief summary of the incident events can
be found at appendix 5. Note that an incident may have involved more than one type of error
(see Table 1).
The most frequently reported type of error was system operated in unsafe condition. This
included incidents where aircraft systems such as flaps or thrust reversers were operated dur-
ing maintenance when obstructions or workers were in the vicinity. In some cases aircraft sys-
tems were operated while the system was partly disassembled. The following incident is an
example of a system operated in an unsafe condition:
Table 1
Frequency of Maintenance Error Types
Type of Error
System operated in unsafe condition 16
Towing event 10
System not made safe 10
Equipment failure 10
Degradation not found 6
Falls and spontaneous actions 6
Incomplete installation 5
Work not documented 5
Person entered dangerous area 5
Person contacted hazard 4
System not reactivated/deactivated 4
Did not obtain or use appropriate equip. 4
Unserviceable equipment used 4
Verbal warning not given 3
Vehicle driving (not towing) 2
Pin or tie left in place 2
Warning sign or tag not used 2
Not properly tested 2
Safety lock or warning removed 2
Vehicle/equipment contacted aircraft 2
Material left in aircraft/engine 1
Access panel not closed 1
Contamination of open system 1
Equipment not installed 1
Panel installed incorrectly 1
Required servicing not performed 1
Unable to access part in stores 1
Wrong equipment/part installed 1
Wrong fluid type 1
Wrong orientation 1
Unable to be coded 6
8
On a night shift at about 2.30 a.m., the crew were performing a rigging check on the
leading edge slats after gearboxes had been changed. At the end of this work, one
torque tube was left disconnected from a gearbox. The crew went for a break and after
they returned they extended the slats on hydraulics as a check of their work. One sec-
tion of the slats did not move, because of the unconnected torque tube. As a result part
of the leading edge was torn off. The crew would normally have checked their work
before extending the slats, but this check had been omitted because of the interrup-
tion caused by the break. The reporter considered that in hindsight, it would have been
more sensible to have extended the slats electrically as they would have moved more
slowly. In addition, had docking been available, they could have walked along the
length of the wing to check the connections.
The second most frequently reported error reflects the potential for damage to aircraft as they
are manoeuvred by maintenance personnel in areas where space is restricted. The following is
an example of such an incident:
As a large multi-engine aircraft was being pushed out of a hangar on a dark night, the
left winglet hit the left stabiliser of a twin-engine aircraft. An engineer stationed under
the wing saw that the collision was about to occur and made torch signals to alert the
licensed engineer stationed near the nose of the aircraft, who could have given the tug
driver a signal to stop the tug. However, the licensed engineer did not see the torch
signals in time to prevent the collision. The twin-engine aircraft had to be re-scheduled
while the damage was repaired. The reporter stated that he was not aware of any stan-
dard procedures for stopping aircraft with torch signals. The reporter considered that
fatigue was a factor in this occurrence.
The third most frequent error, system not made safe, refers to situations where an aircraft sys-
tem was not disabled or locked out appropriately before work commenced. Included are
instances where electrical power was left on while electrical work was carried out and instances
where hydraulically activated systems were not isolated from hydraulic power.
Equipment failure refers to situations where an item of maintenance equipment or an aircraft
component failed and this was not a result of maintenance actions. On some occasions, an
equipment failure combined with a human error to create the incident; for example, there were
two occasions where workers’ unsafe behaviour brought them into contact with faulty electri-
cal equipment, resulting in non-fatal electric shocks.
Incident outcome
Incidents were coded using the coding sheet reproduced in appendix 2. The end results of the
incidents are summarised in Figure 4.
Figure 4 Outcome of maintenance incidents.
Number of incidents
0 5 10 15 20 25 30 35
Potential hazard
Exposure to hazard
Damage to aircraft
Correction of problem
Delayed aircraft
9
Incidents which related to the safety of workers (OH&S incidents) were classified according to
whether they resulted in death of a worker, exposure of the worker to a hazard or potential
exposure of the worker to a hazard. None of the reported incidents resulted in death. Incidents
with airworthiness implications were classified according to whether they resulted in damage
to an aircraft, potential damage to an aircraft, an aircraft signed off with an unrectified prob-
lem, a delayed aircraft, an aircraft signed off with a problem which resulted from maintenance
action, or the detection and correction of a problem.
Note that nine incidents involved both an airworthiness element and an OH&S risk. In Figure 4,
these incidents have been counted separately under both the appropriate airworthiness and
OH&S category. As can be seen, the most frequently reported incident outcome was potential
hazard, where there was a risk that a worker could have been exposed to a hazard such as
hydraulically activated aircraft components or dangerous working surfaces. Exposure to hazard
refers to situations where the final outcome of the incident was that a worker came into contact
with a hazard, whether or not they had any control over the situation. Two examples of this inci-
dent type are a worker who was doused in fuel and a worker who received a cut hand when he
came into contact with windmilling engine fan blades. Correction of problem refers to situations
where a maintenance error was made but then recognised and corrected before the work was
signed off. For example, a part was installed upside down, but then removed and reinstalled cor-
rectly by the same workers. Potential damage to aircraft includes situations where an aircraft sys-
tem was not disabled before maintenance work was carried out and where the system would
have been damaged if it had been activated during maintenance.
Reporting of incidents
Most incidents which had an airworthiness element had been officially reported within the
company; however, only a minority of the incidents with OH&S implications had been offi-
cially reported (see Figure 5). In considering the unreported airworthiness incidents, it should
be noted that it would not normally have been necessary to report an incident which resulted
in only potential damage to an aircraft or where a mistake was rectified.
30
Number of incidents
25
20
15
10
0
Reported Not Information
previously reported not available
previously
Airworthiness OH&S
10
Figure 6 Previous occurrence of similar incidents
25
20
Number of incidents
15
10
0
Has happened Has not Information
before happened before not available
Airworthiness OH&S
Time of incident
The times of day at which incidents occurred is summarised in Figure 8. There were two peaks,
one in the late morning after 1000 hours, the other in the mid-evening between 1800 and 2000.
Care should be exercised in interpreting these figures as they may reflect the hours at which
most maintenance work is carried out and may not necessarily indicate a higher rate of inci-
dents in a particular time period.
As there were no discernible differences between the time at which OH&S and airworthiness
incidents occurred, and because of the relatively small number of incidents in each 2-hour time
period, information relating to the time of occurrence is not broken down by incident type.
At the time that the study was conducted, the day shift commenced at 0600 and concluded at
1400, the afternoon shift commenced at 1400 and concluded at 2200, and the night shift com-
menced at 2200 and concluded at 0600. There were approximately three times as many work-
ers present during morning and day shifts as during night shifts. Hence, if the frequency of
incidents reflected only the number of workers present and not the time of day, it would be
expected that there would be three times as many incidents during morning and day shifts as
during night shifts.
Figure 9 presents the relative number of incidents at each time of day when the variation in the
number of workers present has been taken into account by dividing the incidents which
occurred during morning and day shifts by a factor of three.
In essence, this figure presents the relative frequency of incidents which could be expected if the
same number of workers were present at all times throughout the day. A cyclic pattern of inci-
dents can be seen, corresponding to the three shift patterns described above. However, the
greatest relative frequency of incidents occurred on the night shift between the hours of 0200
and 0400.
11
Figure 7 Area of aircraft involved in the incident
Area of aircraft
Engine
Flight controls
Thrust reversers
Wings
Landing gear
Galley
Fuselage
Hydraulics Airworthiness
incidents
Fuel
OH&S incidents
Engine oil
Doors
Stabiliser
Radar
Pneumatics
Igniters
Fuel pumps
Engine strut
Engine air
Electrical power
APU
Air-conditioning
0 2 4 6 8 10 12 14
Number of incidents
10
8
Number of incidents
0
0
0
0
0
0
0
0
0
0
0
00
0
60
20
80
40
20
80
40
00
60
40
00
02
-1
-1
-0
-0
-2
-1
-1
-1
-0
-2
-2
-
01
01
01
01
01
01
01
01
01
01
00
01
14
10
06
02
20
16
12
08
04
22
00
18
Time of day
12
Figure 9 Relative frequency of incident corrected for number of workers present
4.0
2.5
2.0
1.5
1.0
0.5
0.0
0
0
0
0
0
0
0
0
0
0
0
20
60
80
40
20
80
40
00
60
40
00
20
-1
-1
-0
-0
-2
-1
-1
-1
-0
-2
-2
-0
01
01
01
01
01
01
01
01
01
01
01
00
10
14
06
02
20
16
12
08
04
22
18
00
Time of day
15
12
Number of incidents
0
1 2 3 4 5 6 7 8
Incident involvement
An incident may have involved more than one event. For example, a worker may have made an
error during the installation of a component and a second worker may then have failed to
detect the error during a check. Each event in the incident sequence was classified according to
whether it involved the actions of a person, a failure of equipment or an event in the environ-
ment (such as a gust of wind).
13
Eighty-seven per cent of the incident events involved the actions of people. Equipment failures
accounted for 12% of events, while environmental events represented less than 1% of the total
events (see Figure 11).
100
90
80
70
% of events
60
50
40
30
20
10
0
People Equipment Environment
Event type
Human errors
Errors were categorised with Swain’s error classification system, described earlier.
Omissions constituted the greatest proportion of human errors in the incidents (see Figure 12).
Commissions constituted the second-most frequent type of error.
Mis-timed actions
3%
Substitutions
12%
Omissions
48%
Commissions
37%
14
Analysis of Incidents
In the following section, the incidents are analysed using models of human error and accident
causation. Maintenance errors are first analysed with the skill-rule-knowledge framework
developed by Rasmussen. The Reason model is then applied to the incidents, first by examin-
ing the local factors which contributed to the incidents and then by considering the wider
organisational factors which were involved in incidents. This section concludes with a qualita-
tive analysis guided by the SHEL model described in the introduction of this report.
Unclassifiable
10%
Probably knowledge Probably skill
4% 23%
Skill
11%
Rule
34%
Probably rule
18%
Knowledge-based errors were relatively rare and only 4% of errors were classified as ‘probably’
knowledge-based errors. Thirty-four per cent of the errors were skill-based. There was a degree
of uncertainty in classifying these errors, and most of these errors could only be classified as
skill-based errors with less than complete certainty. An example of a skill-based error follows.
A technician working under the fan cowl of an engine left a large spanner wedged
between tubing. After the aircraft had departed, the technician realised that the span-
ner was missing but did not take steps to alert the airline. The technician retrieved his
spanner when the aircraft returned.
The initial error in this case was a skill-based lapse, related to an inadvisable work practice
(resting tools on convenient parts of aircraft). The technician’s subsequent response to this
lapse can be seen as a rule-based mistake.
In 52% of cases, the action took the form of a rule-based error. For example:
An ignition check was being performed on a GE CF6 engine using a test box. Both
ignition plugs were disconnected and one igniter at a time was plugged into the box
where it was fired. The wrong cannon plug at the low tension end had been discon-
nected, with the result that the igniter which had been left hanging free in the air was
still connected. During the igniter check, the loose igniter fired and sparked across to
the engine. The person who disconnected the cannon plug had thought (wrongly) that
15
the upper ignition box connected to upper ignition. But in fact in the CF6 engine, the
top ignition box provides power to the bottom ignition and vice versa. The reporter
conceded that it had been a bad maintenance practice to leave both leads out at the
same time.
The person involved made a rule-based mistake by assuming that disconnecting the top can-
non plug would deactivate the top igniter.
0 3 6 9 12 15
Number of events
16
Figure 15 Most frequent local factors in airworthiness events
Factor
Procedures
Communications
Control and supervision of work
Preceived pressure or haste
Knowledge, skills, experience
Tools and equipment
Space restrictions
Aircraft design
Documentation
Shift change
Fatigue
Distraction
Environment
0 2 4 6 8 10 12 14 16 18 20
Number of events
Some factors were reported less frequently than anticipated. For example, fatigue was not one
of the more frequent factors. When it was reported, the fatigue had sometimes been induced
by social life outside work rather than the work itself. Anecdotal evidence suggests that task
interruption is a significant challenge to work quality, yet this factor was mentioned in only a
few cases.
Equipment maintenance
System defences
Design issues
Materials
Management oversight
OH&S
Communications
Norm/procedure conflict
Training
0 5 10 15 20
Number of events
17
Control of procedures emerged as the most frequent organisational factor. A local difficulty with
the application of procedures may reflect a wider organisational problem in the development
and dissemination of procedures. Hence, procedures feature as both local and organisational
factors. These results suggest that management may need to take a more active role in standar-
dising and documenting procedures, and in ensuring that procedures are followed.
Equipment maintenance emerged as an organisational level factor relevant to both OH&S and
airworthiness incidents. This is closely related to the incidence of equipment deficiencies as a
local factor, in that the presence of broken or faulty equipment (such as stands and lighting) on
the hangar floor may reflect a wider organisational problem in the maintenance of equipment.
System defences generally refers to procedures such as engine runs and dual inspections intend-
ed to ensure that maintenance has been carried out correctly. On some occasions where this
factor was identified, incorrect maintenance was not detected and corrected before the aircraft
was released from maintenance.
18
One maintenance crew was working on the left engine of the aircraft. Another crew
working on the right engine had locked out the leading edge flaps, and had reverser
cowls open but had not locked out reversers. If the reversers had been activated, it
would have severely damaged the reverser halves. The crew working on the left-hand
engine believed that procedures required the reversers to be locked out and docu-
mented. A disagreement ensued between members of the two crews about whether it
was necessary to lock out the reversers.
In another case, a crew reported that they performed additional functional checks on aircraft
systems, not called for in procedures, as they believed that the procedures were inadequate.
While the additional checks may be beneficial, they result in a variation in procedures across
crews.
Some small tasks, such as activating hydraulics or moving flight controls appear to be per-
formed according to the individual work habits and common sense of the engineer rather than
in accord with formal procedures or guidelines. This informal aspect of task performance
increases the chance that steps will be omitted, that systems will be activated in unsafe condi-
tions, or that tools or other devices will be left inside aircraft. A greater use of formal checklists
or procedures could help to avoid some of the incidents which were related to lapses or incor-
rect assumptions about system status. It should also be possible to periodically review the work
practices of individuals and crews and ensure that divergent work practices do not develop and
that procedures are understood.
Training
There was a view among some interviewees that recurrent training was lacking. This included
refresher training and training in some new systems. It was reported that some personnel had
serviced systems such as the Traffic Alert and Collision Avoidance System (TCAS) without
appropriate training.
There were also indications that the training system does not always receive adequate feedback
on the problems which occur on the hangar floor. For example, some recurring problems could
be addressed by changes to the training system.
Defences
Authorities on human factors in aviation such as Hawkins (1993) recommend a ‘two-pronged’
approach to human error. First, errors should be minimised. Second, the consequences of those
errors which nevertheless occur should be reduced. Defences are part of this second approach
to error. Defences can be built into the system to catch errors before they have the opportuni-
ty to cause serious consequences.
Examples of defences in maintenance are functional checks, dual certification and visual
inspection of systems. Several incidents indicated that defence ‘safety nets’ were failing to catch
maintenance errors which had occurred in earlier maintenance procedures. The following inci-
dent involved a defence which failed to identify a maintenance error.
19
A fuel filter was being re-fitted to a GE CF6 engine after a fuel pump had been
changed. Six studs held the filter on, with two in an inaccessible position. It was not
possible to see the nuts on two of the studs without a mirror. These two nuts were left
off, but the other four had been installed correctly. Two workers were performing the
task, an apprentice and a licensed aircraft maintenance engineer (LAME). They knew
there were six nuts to install, but thought they had fitted them all. There were many
loose parts around. A leak check at idle power didn’t reveal any leaks. When the air-
craft took off, the pilot was advised by air traffic control that there was vapour stream-
ing from the engine. The aircraft continued to its destination, and on landing, the pilot
was again informed of vapour by the tower. A power run was carried out at the desti-
nation and the problem was identified. It was subsequently determined that more than
20 such incidents had occurred worldwide at various airlines.
In this case the defence, a leak check, failed to disclose a problem which subsequently became
evident at takeoff. This was not the only occasion in which an idle power run or engine spin
failed to show up a problem.
Shift handover
Overseas accident experience has indicated that inadequate shift handover can be a significant
problem for maintenance organisations. However, shift handovers were attributed as a factor
in fewer than 10% of the current airworthiness incidents. No information is available on the
proportion of jobs which require more than one shift to complete. If such information were
available, it would be possible to express shift handover problems as a proportion of all shift
handovers.
The following relatively minor incident was related to a shift change:
During a routine engine change during night shift, a strut inspection was performed
after engine removal and it was noticed that there was a crack in the pre-cooler. The
fault was marked on the pre-cooler and was documented. A new pre-cooler was
ordered and was delivered and put under the engine before the night shift went home.
The day shift arrived and in their eagerness to get the job done, installed the new
engine without changing the pre-cooler. This was partly understandable as the work
schedule did not include the pre-cooler and the crew did not browse through the doc-
umentation sheets before starting work.
Communication
Sometimes the dissemination of information within maintenance organisations is inadequate.
For example, at times a new task such as a modification led to mistakes when a crew was tasked
with performing the work for the first time. Other crews which had previously performed the
task had also made the same mistakes, yet the lessons had not been communicated to all appro-
priate personnel.
Communication between technicians with differing trade backgrounds can also be a problem.
This was evidenced by several incidents which involved misunderstandings between specialist
technicians such as personnel with electrical qualifications and those with engine and airframe
licences.
Equipment issues—hardware
The hardware element of the SHEL model refers to the physical equipment used in the job.
Several issues emerged concerning this aspect of the work of maintenance engineers, particu-
larly in incidents with OH&S implications.
Equipment maintenance
Problems with the maintenance of hangar equipment such as stands, lighting system and vehi-
cles featured frequently in incident reports. In some cases, equipment which had been identi-
fied as unserviceable was still available for use. For example:
While changing the logo light in the tail of an aircraft, the lift truck, which had been
extended over the tail (horizontal stabiliser), slowly dropped onto the stabiliser. The truck
had been snagged as unserviceable due to a hydraulic leak, but had still been issued
to the workers. When the truck platform was retracted, the main supports gouged into
the leading edge of the stabiliser. The worker reported that he had learnt from this expe-
rience to avoid, where possible, putting ground equipment over an aircraft.
Although some equipment has been in service for many years, several reporters specifically
mentioned that some older equipment, particularly old stands and mobile stairs, were still use-
ful for many maintenance tasks.
Design issues
There were several incidents where an error made by engineers was related to the design of air-
craft or engines. For example, several incidents were reported in which thrust reversers were
left locked out on twin-engine aircraft (an example of such an incident has been reported ear-
lier). All of these incidents involved aircraft equipped with GE CF6 engines: none occurred on
Pratt and Whitney engines. These incidents appear to be directly related to the design of the
lockout system. Other incidents involving the incorrect or incomplete assembly of components
appear to reflect designs which increase the potential for error. The example of an incomplete
installation of a fuel filter referred to earlier illustrates this problem.
22
It would appear that ergonomic maintenance considerations are sometimes overlooked when
aircraft and systems are being designed. Better feedback to manufacturers on the errors which
occur during the maintenance of their products may help to reduce design-related mainte-
nance errors.
23
Discussion
The data collected in this project has provided some useful insights into the nature of mainte-
nance incidents.
Most types of incidents were repetitive. The recurring nature of such incidents means that they
are to some extent predictable, and hence measures can be taken to prevent them.
Many incidents involved flight controls, engines, and thrust reverser systems.
Human error featured in most incidents, with omissions being the most frequent type of error.
The skill-rule-knowledge framework proved to be an adequate classification system for main-
tenance errors, although in some cases, the framework was difficult to apply. Errors due to
inadequate knowledge were rare and were usually committed by trainee technicians. Skill-
based errors occurred in approximately 34% of the incidents. However, the majority (52%) of
the errors took the form of rule-based mistakes. Many of these errors took the form of mistak-
en assumptions, particularly where workers wrongly assumed that an aircraft system was in a
particular configuration.
Williamson and Feyer (1990), in examining the errors which preceded a sample of 1,020 work-
related fatalities, found that just over 50% of errors were skill-based, while fewer than 14% were
rule-based mistakes and fewer than 14% were knowledge-based mistakes. In contrast, the cur-
rent study, while based on a significantly smaller number of incidents, would appear to indi-
cate that formal and informal rules play a significant part in guiding the actions of maintenance
technicians and that these rules provide a significant source of errors.
A common rule-based error was the activation of an aircraft system (such as hydraulics), with-
out first checking the status of cockpit controls. Remedial action for such errors could include
ensuring that technicians acquire appropriate situational awareness before activating systems,
possibly by the use of checklists to guide performance of routine actions.
The type of error which can be expected on a given task appears to be closely related to the fre-
quency with which that task is performed. By definition, rule-based or knowledge-based mis-
takes can be expected to be particularly prevalent when the task is unusual, but to become less
common on tasks which are performed frequently. Skill-based slips and lapses, on the other
hand, are relatively uncommon on unusual tasks but may become problematic on tasks which
are performed routinely.
For example, frequently performed tasks such as routine boroscope inspections have the poten-
tial to produce skill-based errors because of the familiarity of the work and the potential for
absent-minded task performance. An important consequence of the relationship illustrated in
Figure 17 is that by categorising tasks according to the frequency of their performance, main-
Figure 17 Task frequency and the prevelance of error
24
tenance managers will be better placed to anticipate errors and introduce appropriate counter-
measures. A further implication of this relationship is that for individual workers, there may be
an optimum level of task familiarity lying somewhere between the extremes of highly unusual
and highly routine tasks.
Several patterns emerged in the timing of incidents. In general, the frequency of incidents grad-
ually increased as the shift progressed, with a peak of incidents in the second-last hour of the
shift. The last hour of the shift, however, was the time when the least number of incidents
occurred. This may reflect a change in the work being carried out, as workers clean up work
areas and prepare to finish work for the day.
Research on the timing of workplace and transport accidents indicates that the early hours of
the morning and early afternoon are times at which accidents are particularly likely to occur.
For example, Williamson and Feyer (1995) found that when the number of people at work
throughout the 24 hours of the day is taken into account, fatal workplace accidents in Australia
are more likely to occur at night than during the day. The results of this study are consistent
with those findings, and indicate that the rate of incidents per worker is greatest in the period
between 0200 and 0400.
The local factors underlying airworthiness incidents were generally different to those underly-
ing OH&S incidents. For airworthiness incidents, procedures, control and supervision and
communications emerged as the most significant factors. For OH&S incidents, tools and
equipment, perceived pressure or haste, and the physical environment were the three most fre-
quent local factors.
Attributing local factors was relatively straightforward in most cases. However, organisational
factors could not be identified with a high level of reliability using the coding sheet presented
in the appendix.
Much of this document to this point has been concerned with the factors which lead to errors.
However, it is important to recognise that errors can also be addressed in at least two other
ways.
First, some errors occur while preventative maintenance or inspection tasks are being per-
formed. It is possible that in some cases, the risk of a system failure caused by a maintenance
error will be greater than the risk of a failure if the system is left undisturbed. A well known
example which illustrates this problem occurred in North America when an L1011 lost oil from
all three engines when magnetic chip detectors were fitted without O-rings. The benefits of the
inspection program would certainly have been outweighed by the risk of such an error occur-
ring (National Transportation Safety Board 1984).
In planning maintenance tasks, it may be necessary to factor in the probability of a human
error occurring during the task. For example, an inspection task which requires the frequent
opening and closing of access covers is likely to produce occasional skill-based lapses where
access covers will be left unsecured. Where the maintenance is designed to achieve financial
rather than safety benefits, such as extending the service life of a system, there may be occasions
where it would be preferable to leave a system undisturbed in order to avoid the possibility of
a maintenance error.
The second approach to human error is to acknowledge that errors will occur from time to
time and to design procedures and systems which can tolerate such errors. Avoiding the simul-
taneous performance of the same task on similar systems is an example of such an approach.
For example, on 25 February 1995, a European-operated 737-400 was forced to divert shortly
after departure following a loss of oil quantity and pressure on both engines. Both of the air-
craft’s engines had been subject to boroscope inspections during the night prior to the incident
flight. High-pressure rotor drive covers had been left unsecured on each engine and, as a result,
nearly all the oil had been lost from each engine during the brief flight (Air Accidents
Investigation Branch 1996).
Where extended range twin-engine operations (ETOPS) are being conducted, the performance
of identical maintenance actions on multiple elements of critical systems is avoided wherever
possible. Engines, fuel systems, fire suppression systems and electrical power are examples of
25
ETOPS critical systems on aircraft such as the B767 and B737. Boeing lists several approaches
designed to minimise the impact of human error in maintenance of ETOPS aircraft. These
include performing maintenance actions on different legs of a flight, having two identical tasks
performed by different mechanics, adding inspections to detect errors, and carrying out tests
to verify that maintenance has not introduced a problem (Boeing 1994a). However, these pre-
cautions to minimise the impact of human error are not generally applied to aircraft with more
than two engines, or to twin-engine aircraft which are not being maintained in accordance with
an ETOPS maintenance program.
The extension of ETOPS philosophies to non-ETOPS aircraft would help to contain mainte-
nance induced problems. For example, staggered maintenance may reduce the risks associated
with simultaneous maintenance of similar systems.
26
Conclusion
It should be noted that this study is based on a limited sample of maintenance incidents that
may not be representative of all possible maintenance incidents. Hence, the issues raised in this
report will not necessarily apply to all maintenance operations. Furthermore, much of the
information in this report is derived from the knowledge and experience of maintenance engi-
neers themselves and the results will to some extent reflect their perceptions. Nevertheless, a
picture has emerged of the types of incidents which occur during maintenance work and the
factors which lead to these problems. The airworthiness incidents examined in this report indi-
cate that procedures, equipment, communications and control, and supervision of work may
be worthwhile areas for attention.
27
Safety Actions
The following suggested safety actions are presented as a result of this study. Organisations will
need to assess the extent to which these suggested safety actions are appropriate for their own
operations.
1. Procedures
1.1 Maintenance organisations should periodically review documented maintenance proce-
dures to ensure that they are accessible, consistent and realistic.
1.2 Maintenance organisations should periodically examine work practices to ensure that
they have not evolved to the extent that they differ significantly from formal procedures.
Narrowing the gap between work practices and formal procedures may require modifi-
cations to procedures as well as to work practices.
1.3 Maintenance organisations should, where possible, ensure that standardised work prac-
tices are adhered to across their maintenance operations.
1.4 Maintenance organisations should evaluate the ability of checklists, whether performed
from memory or from paper, to assist the performance of maintenance personnel in rou-
tine situations such as activating hydraulics, moving flight surfaces or preparing an air-
craft for towing.
3. Communication
3.1 Maintenance organisations should ensure that adequate systems are in place to dissemi-
nate important information to all maintenance personnel, particularly where procedures
have changed or where an error has occurred repeatedly on a task.
5. Shift handover
5.1 Maintenance organisations should review the adequacy of shift handover practices, with
particular attention to documentation and communication, to ensure that incomplete
tasks are seamlessly transferred across shifts.
28
6. Supervision
6.1 Maintenance organisations should recognise that supervision and management oversight
may need to be increased, particularly in the last few hours of each shift, as errors become
more likely.
7. Towing aircraft
7.1 The procedures and equipment used to tow aircraft to and from maintenance facilities
should be reviewed. Particular attention should be paid to the possible need for a direct
verbal communication link between engineering personnel and tug drivers, to reduce the
current heavy reliance on hand signals. The procedures used to maintain safe clearance
between towed aircraft and obstructions in confined spaces should also receive attention.
8. Design issues
8.1 Manufacturers should give greater consideration to maintenance ergonomic issues when
designing systems and should actively seek information on the errors which occur when
systems are being maintained.
9. Training
9.1 Maintenance organisations should consider introducing crew resource management
training for maintenance engineers and other personnel (such as tug drivers) who inter-
act with maintenance personnel.
9.2 Regular refresher training should be offered to maintenance engineers with particular
emphasis on company procedures. Such training could help to reduce the frequency of
incidents related to misunderstandings of company procedures.
29
Appendix 1
Maintenance Incidents
I am studying incidents in aircraft maintenance. By maintenance incident, I mean any situa-
tion in which something happened which could have prevented the aircraft from operating
normally or could have put the safety of anyone (including maintenance workers) at risk. I am
interested in any occasion when a problem happened in maintenance, including the big prob-
lems and the small mistakes and including the times when a problem occurred but was cor-
rected before the aircraft was signed back to the line.
I know that most of the time, maintenance is uneventful, but by gathering information about
the occasions when things go wrong, I hope to learn about the entire maintenance system.
A. First, I would like you to describe an incident that put at risk your safety, or the safety of
one of your workmates. I am only interested in incidents that happened in the last 12
months or so, that you actually saw happen, either to you or to someone else. Your iden-
tity will remain confidential and if you want, the information you give me will also
remain confidential. I am interested in the incident itself, and what it tells us about the
maintenance system. I am not interested in individuals and I am not interested in blam-
ing people for things that have happened.
30
8. Was it a routine task?
9. Did the incident occur while a problem was being dealt with?
–People
–Equipment
–Environment
–Task
13. Would you object if I included this incident in a report for Unions and Management?
B. Now can you tell me about a maintenance incident that could have prevented an aircraft
from operating normally. Again, I am interested in incidents which occurred in the last
12 months or so, that you actually saw happen, either to you or to someone else.
31
3. Was it officially reported? Was corrective action recorded?
4. Could it happen again?
5. What time was it?
6. How long into the shift was it?
7. Did it involve another shift/another crew?
8. Was it a routine task?
–was the person distracted?
–would their actions have been correct in other circumstances?
9. Did the incident occur while a problem was being dealt with?
–was there a standard set of rules to deal with this problem?
10. What type of work was being performed at the time?
11. What part of the aircraft was being worked on?
12. Why do you think the incident happened?
–People
–Equipment
–Environment
–Task
13. Would you object if I included this incident in a report for Unions and Management?
32
Appendix 2
Human factors in maintenance coding sheet
1. Case number___ Coder___ Reporter___ Aircraft type______
2. Type of Incident
33
(16.4) Description of event 4
13. What part of the aircraft was involved? (ATA Chapter) _________
14. Would the reporter object if this incident was made public? YN
15. EVENT 1
(A) Yes
(B) Probably
(C) Probably not
(D) No
(E) Don’t Know
(15.2.5)
If behaviour was abnormal and/or rule breaking, was action:
(A) Omission Yes Probably
(B) Commission Yes Probably
(C) Substitution Yes Probably
(D) Mis-timed action Yes Probably
(E) Unclassifiable in above categories Yes
1 PHYSICAL FACTORS
1A Anthropometric characteristics
1B Sensory limitations
1C Fatigue
1D Drugs (including alcohol)
2 COGNITIVE FACTORS
2A Distraction from task
2B Memory
2C Workload
2D Knowledge, skills and experience
2E Perceived pressure or haste
2F Convenience
3 BETWEEN PEOPLE
3A Planning within crew
3B Visual signals or visibility of other workers
3C Communication
3D Control and supervision of work
4 TASK ASPECTS
4A Shift change
4B Procedures
4C Documentation
4D Task order within work package
35
4E Interruption
4F Conflict between norms and formal procedures
4G Job scheduling
4H End of shift
5H Overtime
5 PHYSICAL OBJECTS
5A Inadequate tools and equipment
5B Space restrictions
5C Illumination
5D Design of aircraft component or system
5E Automated systems
6 PHYSICAL ENVIRONMENT
Includes:
(Temperature extreme, glare, noise, air quality, weather, darkness, heights, working surface or
ground surface)
16.2 Organisational factors relating to event 1
(A) Incompatible goals
(B) Inappropriate structure
(C) Poor communications
(D) Poor planning
(E) Management oversight, control and monitoring
(F) Design failures
(G) Inadequate system defences
(H) Unsuitable materials
(I) Control of procedures
(J) Poor training
(K) Inadequate maintenance of equipment
(L) Inadequate regulation
(M) Conflict between norms and formal procedures
36
Appendix 3
Reliability of coding
For event types and skill rule knowledge error types, the reliability of coding was assessed by
calculating values of Cohen’s kappa (Breakwell, Hammond & Fife-Schaw 1995). Cohen’s kappa
ranges between 0 and 1 and represents the proportion of agreement corrected for chance. Fleiss
(1981) describes values of Cohen’s kappa greater than 0.75 as ‘excellent’, and values between 0.6
and 0.75 as ‘good’ (reported in O’Hare and others 1994).
Forty incidents were coded by a second coder, in addition to the coding conducted by the pri-
mary coder. The forty incidents contained a total of 63 incident events.
Incident factors
Local factors
For the purposes of calculating the reliability of factor codings, local factors were aggregated
into the broad category headings as seen on the coding sheet elsewhere in the appendix. Local
factors assigned to events by the primary coder were also identified by the second coder on 65%
of occasions.
Organisational factors
The reliability of organisational factors was significantly less than that for local factors, and
only 33% of organisational factors identified by the primary coder were identified by the sec-
ond coder.
37
Appendix 4
Definitions
Communications. Information which is essential for the safe functioning of the organisation
does not reach the necessary recipients. Includes communication from the shop floor upwards
and from management downwards.
Commission. An error which occurs when an operator performs a task incorrectly or performs
a task which is not appropriate to the circumstances.
Conflict between norms and formal procedures. Actual work practices do not reflect formal
procedures. In some cases this may because formal procedures are cumbersome or difficult to
comply with.
Control of procedures. Important safety, working or operational regulations and procedures
are not clear, difficult to follow, incorrect, incomplete or inaccessible.
Design failures. Relates to the design of tools, equipment, work environments and procedures.
Inadequate design may promote errors or violations or may produce situations where non-
standard performance results in negative and irreversible consequences.
Environmental events. Relates to elements of the physical environment in which the task was
being performed which cannot be changed or compensated for. Weather-related phenomena
are examples of environmental events.
Equipment-related events. Concerns breakages or malfunctions of equipment, tools and
machinery, including aircraft components.
Event. An occurrence, action or closely related series of actions which occurred at a particular
place and time. Where the actions are behavioural, and a series of actions were performed by
different people, separate events are required for each person.
Inadequate maintenance of equipment. Deficient management of maintenance of tools and
equipment, including lack of preventative maintenance, poor maintenance scheduling and
excessive delays in repairing equipment.
Inadequate regulation. Inadequate surveillance by government bodies or their delegated
authorities. Absence of appropriate laws to regulate dangerous operations.
Inadequate system defences. An absence or inadequacy of ‘safety nets’ in the system. Such
defences can serve to detect or prevent unsafe job performance, or minimise the consequences
of such performance. Includes cases where existing defences are circumvented.
Inappropriate structure. Relates to the structure of the organisation. For example, manage-
ment responsibilities may be blurred or poorly defined.
Incident. A situation in which events occurred which could have prevented an aircraft from
operating normally or could have put the safety of any person (including maintenance work-
ers) at risk.
Incompatible goals. Organisations are generally pursuing several goals at the one time, for
example safety and production. This factor should be coded in cases where the incident events
reflect a poorly resolved goal conflict.
Knowledge-based errors. Occur when a person is required to solve an unfamiliar problem for
which no procedures exist. Knowledge-based errors are typically related to an inappropriate
mental model of the system or a lack of resources for dealing with a complex problem. Rule-
based errors and knowledge-based errors are commonly referred to as mistakes.
Management oversight, control and monitoring. Relates to the way management monitors
the performance of planned work and ensures that work is completed in accord with guide-
lines.
Planning. Where management planning and scheduling of activities has not adequately taken
safety considerations into account.
38
Poor training. Indicated by deficiencies in the knowledge or skills of employees.
Rule-based errors. Occur in familiar situations where, in the process of dealing with a situa-
tion, the worker either applies a bad rule or misapplies a good rule. A typical case is the appli-
cation of a general procedure to a specific situation which calls for a modified version of the
procedure .
Rule-based errors and knowledge-based errors are commonly referred to as mistakes.
Skill-based errors. Occur in rapid automatic mode where the person has invested a minimum
of mental resources to the task at hand. Such errors occur in familiar situations involving
highly practiced automatic routines which may have developed after extensive training and
experience.
Two particular types of skill-based errors are slips and lapses. Slips are errors in which an oper-
ator intends to perform a correct action but accidentally performs another well-learnt action
or action sequence. A lapse is a failure to carry out an intended action. An example is the omis-
sion of a step in a procedure following an interruption.
Substitution. An error in which an operator performs an action in place of the desired action.
Substitution can be considered to be a special type of commission error.
Unsuitable materials. To be coded where the event is related to the provision of unsuitable
tools and equipment or other work materials.
39
Appendix 5
Maintenance incident events
Note: This listing of incident events excludes six incidents where the reporter asked that the inci-
dent remain confidential.
Event description
—Thrust reversers locked out during A check, not documented.
—Thrust reversers not reactivated.
—Worker knowingly used unserviceable stand.
—Acting leading hand suggested that worker remain in engine during towing.
—Worker remained in engine as the aircraft was towed.
—During work on leading edge slats one torque tube not connected.
—Slats extended with one torque tube not connected.
—Stair locking mechanism failed.
—Worker on stairs reacted inappropriately to situation by holding onto engine while still
standing on stairs.
—Worker did not take wind into account when depressurising reservoir.
—Wind blew hydraulic fluid into eyes.
—Rigging pin used without attached streamer.
—Rigging pin left in flight controls (found before work signed off).
—Circuit breaker pulled without being tagged before engine spin.
—Worker pushed circuit breaker in to its normal position and fuel spilled.
—Thrust reversers locked out when not strictly necessary.
—Thrust reversers not reactivated before aircraft dispatch.
—Crew ran engines causing blast to affect other workers.
—Tug driver delayed stopping once signal to stop had been given.
—Tug driver delayed stopping once signal given.
—Stairs rolled back as locking system failed to hold.
—Workers on stairs reacted inappropriately by holding onto aircraft while still standing on
stairs.
—Tradesperson pressed hydraulic switch button while it was tagged as not to be used.
—During a fuel pump change a fuel line was not reconnected properly.
—Lead hand did not check that everyone was clear before calling for a wet spin of engine.
—During inspection of the engine serious damage to part of a fan air valve was missed. (dis-
covered later).
—Incoming shift reinstalled engine without changing pre cooler as documented.
—Igniter circuit breakers not pulled before conducting a wet spin.
—LAME reacted incorrectly to engine start by shutting off fuel and starter, instead of just fuel.
—Driver misjudged tow, aircraft scraped docking.
—Thrust reversers inspected with power running without them being locked in place.
—Incorrectly installed panel on twin-engine aircraft.
40
—Flap lever had been moved up while flaps were in down position with power off.
—Hydraulics had been started without first ensuring that flaps and flap lever were consistent.
—Crew not looking, towed aircraft hit a tree.
—Engine jammed as it was being hoisted.
—Person put fingers between engine and mounts.
—Crew failed to notice that rod end was incorrectly adjusted.
—Previous crew had left aircraft with flap lever inconsistent with flap position and
hydraulics off.
—Worker in cockpit started hydraulics without first confirming that flap lever and flap posi-
tion were consistent.
—Mechanical crew told electrical crew that it was OK to run flaps with panel incompletely fit-
ted, flaps damaged.
—Foot slipped while working on engine.
—Worker did not pull circuit breakers before starting electrical work.
—Tool crib permitted truck to be used which had been snagged US.
—Truck settled on hydraulics while positioned over horizontal stabiliser.
—Platform retracted, scraped aircraft.
—Worker slipped on wet surface.
—Worker attempted to perform task without appropriate equipment.
—Person drove vehicle near engine that was about to be started.
—AME left parked tug with engine running.
—Tug transmission slipped into reverse, collided with engine cowl of 737.
—Cleaner drove vehicle into unsafe area.
—LAME on interphone did not see warning signal of wingman, continued tow until collision.
—Two nuts left off fuel pump filter.
—System did not leak under a leak check.
—Jet blast from B747 affects hangars.
—Worker failed to torque 2 nuts on GE CF6 fuel system.
—Technicians started pneumatics without first warning other workers working on the aircraft.
—Electrical crew failed to lock out reversers when thrust reverse cowls open.
—Workers seen riding on the tray of a utility vehicle, in contravention of safety guidelines.
—Crew failed to re-activate thrust reverse during system re-activation.
—Worker positioned in dangerous location without adequate handholds, resulted in fall.
—Aircraft wingtip struck a tree while being towed.
—No warning PA was made before aircraft was towed, worker was up ladder in cabin at time.
—Crew members sheltered from rain as aircraft pushed back from terminal, insufficient peo-
ple to check clearance.
—Tug driver could not follow line due to poor visibility, aircraft struck parked aircraft.
—Upper deck door was left on automatic after test.
—Door opened automatically, slide deployed.
—Failed to find and correct aircraft fault.
41
—Valve stuck open directing fuel to an area which could have been occupied by a worker.
—Worker did not pull out circuit breaker before working on electrical system, resulted in elec-
tric shock.
—Stand was not moved out of the way before aircraft was towed out of hanger.
—Crew failed to notice that the aircraft was too close to the stand, in time to prevent contact.
—No warning sign used when radar switched on with people in vicinity.
—All four reverse thrust blocker doors were activated while people were working on the air-
craft.
—Wrong valve fitted to aircraft pneumatic system.
—Valve broke—distributing metal fragments in pneumatic system, damaging heat exchangers.
—AME left spanner under engine cowl.
—After aircraft departed, suspected where spanner was, but did not notify company.
—Before compressor wash, blanked off wrong sense line.
—Washed compressor with wrong lines blanked off.
—LAME started APU even though it was tagged for safety reasons.
—Towing cable attached to blast fence badly worn.
—Worker did not ensure that a fire extinguisher was available while conducting work with the
fire system inoperative.
—After adding oil to RB211, LAME replaced oil cap but did not push locking flap down.
—One igniter was left disconnected while other was being checked.
—Wrong igniter was fired, sparked across to engine.
—Worker started working in slippery dangerous area without first taking precautions to reduce
the risk.
—Worker slipped and cut hand on rotating turbine blade.
—Unable to find part (static invertor) in computerised parts inventory system.
—Part was ‘robbed’ from another aircraft.
—Details of ‘robbed’ part were passed to an apprentice instead of being documented in log.
—Apprentice did not act on the information.
—AME accidentally put engine oil into hydraulic system of B747.
—Water heating system (hot cup) failed due to short circuit.
—LAME tested system by placing hand into ‘live’ water, received shock.
—Fault not found on walkaround.
—Earth pin on power lead broken.
—Worker attempted to plug two live leads together.
—Door power assist system malfunctioned when door opened.
—Aircraft signed out of hangar with inoperative spoilers, reason not established.
—Reporter forgot to lower wheels of towbar before tug drove away, towbar dropped on foot.
—Ladder slipped on stand while worker was trying to reach radome.
—Technicians climbed from stand onto an engine.
42
References
Air Accidents Investigation Branch 1996, Report on the Incident to a Boeing 737–400, G-OBMM
near Daventry on 25 February 1995, Aircraft Accident Report 3/96, HMSO, London.
Boeing 1994, Maintenance Error Decision Aid, Boeing Commercial Airplane Group, Seattle.
Boeing 1994a, ETOPS Guide, Maintenance Program Guidelines, Revision A, Boeing Commercial
Airplane Group, Seattle.
Boeing 1996, Statistical Summary of Commercial Jet Aircraft Accidents, Worldwide Operations
1959–1995, Boeing Commercial Airplane Group, Seattle.
Breakwell, G. M., Hammond, S., & Fife-Schaw, C. (eds) 1995, Research Methods in Psychology,
Sage, London.
Civil Aviation Authority (UK), Flight Safety Occurrence Digest 92/D/12.
Dell, G. P., & Ojczjk, U. L. 1996, ‘Aircraft pushback accidents: The myth of carelessness and
common sense dispelled’, Safety in Australia, Mar.
Drury, C. G., & Gramopadhye, A. K. 1990, ‘Training for visual inspection’, in Proceedings of
Third Federal Aviation Administration Meeting on Human Factors Issues in Aircraft Maintenance
and Inspection, Atlantic City, New Jersey, pp. 149–164.
Eitelberg, M. 1991, ‘Increased use of women and minorities in military aviation maintenance’,
in Proceedings of the Fourth Federal Aviation Administration Meeting on Human Factors Issues in
Maintenance and Inspection, Alexandria, Virginia, pp.154–178.
Fitts, P. M., & Jones, R. H. 1947, ‘Analysis of factors contributing to 460 “pilot error” experi-
ences in operating aircraft controls’, in Selected Papers on Human Factors in the Design and Use
of Control Systems, ed. H. W. Sinaiki, Dover, New York.
Flanagan, J. C. 1954, ‘The critical incident technique’, Psychological Bulletin, vol. 51, no. 28,
pp 327–358.
Hawkins, F. 1993, Human Factors in Flight, Ashgate, Aldershot.
International Civil Aviation Organisation 1995, Human Factors in Aircraft Maintenance and
Inspection, Circular 253-AN/151, ICAO, Montreal.
International Civil Aviation Organisation 1993, Human Factors, Management and
Organisation, ICAO Circular 247-AN/148, ICAO, Montreal.
International Civil Aviation Organisation 1992, Ergonomics, ICAO Circular 238-AN/143,
ICAO, Montreal.
Latorella, K. A., & Drury, C.G. 1992, ‘A framework for human reliability in aircraft inspection’,
in Proceedings of the Seventh Federal Aviation Administration Meeting on Human Factors Issues
in Aircraft Maintenance and Inspection, Atlanta, Georgia, pp. 71–81.
Marx, D. A., & Graeber, R. C. 1994, ‘Human error in aircraft maintenance’, in Aviation
Psychology in Practice, Avebury Technical, Aldershot.
Miller, D. P., & Swain, A. D. 1987, ‘Human error and human reliability’, in Handbook of Human
Factors, ed. G. Salvendy, John Wiley & Sons, New York.
National Transportation Safety Board 1984, Eastern Airlines, Lockheed L1011, Aircraft Accident
Report 84/04, NTSB, Washington.
43
O’Hare, D., Wiggens, M., Batt, R., & Morrison, D. 1994, ‘Cognitive failure analysis for aircraft
accident investigation’, Ergonomics, vol. 37, no. 11, pp. 1,855–1,869.
Patel, S., Prabhu, P., & Drury, C. G. 1992, ‘Design of work control cards’, in Proceedings of the
Seventh Federal Aviation Administration Meeting on Human Factors Issues in Aircraft
Maintenance and Inspection, Atlanta, Georgia, pp. 162–172.
Rasmussen, J. 1983, ‘Skills rules and knowledge: Signals, signs and symbols, and other distinc-
tions in human performance models’, IEEE Transactions on Systems, Man and Cybernetics, vol.
SMC-13, no. 3, pp. 257–266.
Reason, J. 1990, Human Error, Cambridge University Press, Cambridge.
Reason, J. 1991, Identifying the latent causes of aircraft accidents before and after the event,
paper presented to the 22nd Annual Seminar of the International Society of Air Safety
Investigators, Canberra, 4–7 Nov.
Runciman, W. B., Sellen, A., Webb, R. K., Williamson, J. A., Currie, M., Morgan, C., & Russell,
W. J. 1993, ‘Errors, incidents and accidents in anaesthetic practice’, Anaesthesia and Intensive
Care, vol. 21, no. 5, pp. 506–519.
Russell, P. D. 1994, ‘Management strategies for accident prevention, Air Asia, vol. 6, pp. 31–41.
Shepherd, W. T. 1991, The FAA human factors program in aircraft maintenance and inspec-
tion, paper presented to Fifth Federal Aviation Administration Meeting on Human Factors Issues
in Aircraft Maintenance and Inspection, Atlanta, 19–20 June.
Shepherd, W. T., & Parker, J. F. 1991, ‘Future availability of aircraft maintenance personnel’, in
Proceedings of the Human Factors Society 35th Annual Meeting, Human Factors and Ergonomics
Society, Santa Monica, pp. 33–36.
Stelly, J,. & Taylor, J. 1992, ‘Crew coordination concepts for maintenance teams’, in Proceedings
of the Seventh Federal Aviation Administration Meeting on Human Factors Issues in Aircraft
Maintenance and Inspection, Atlanta, Georgia, pp. 173–206.
Taggert, W. R. 1990, ‘Introducing CRM into maintenance training’, in Proceedings of Third
Federal Aviation Administration Meeting on Human Factors Issues in Aircraft Maintenance and
Inspection, Atlantic City, New Jersey, pp. 93–110.
Taylor, J. C. 1990, ‘Organisational context in aviation maintenance: Some preliminary findings
and training implications’, in Proceedings of Third Federal Aviation Administration Meeting on
Human Factors Issues in Aircraft Maintenance and Inspection, Atlantic City, New Jersey,
pp 163–176.
Wiener, E. L., Kanki, B. G., & Helmreich, R. L. 1993, Cockpit Resource Management, Academic
Press, London.
Williamson, A., & Feyer, A. 1990, ‘Behavioural epidemiology as a tool for accident research’,
Journal of Occupational Accidents, vol. 12, pp. 207–222.
Williamson, A., & Feyer, A. 1995, ‘Causes of accidents and the time of day’, Work and Stress,
vol. 9, no. 2/3, pp. 158–164.
44