Uploaded by Patrick Moroni Costa

Heart A Proposed Method for Achieving High Reliability in Process Operation by Means of Human Factors Engineering Technology

advertisement
Safety and Reliability
ISSN: 0961-7353 (Print) 2469-4126 (Online) Journal homepage: https://www.tandfonline.com/loi/tsar20
Heart—A Proposed Method for Achieving High
Reliability in Process Operation by Means of
Human Factors Engineering Technology
J. C. Williams
To cite this article: J. C. Williams (2015) Heart—A Proposed Method for Achieving High
Reliability in Process Operation by Means of Human Factors Engineering Technology, Safety
and Reliability, 35:3, 5-25, DOI: 10.1080/09617353.2015.11691046
To link to this article: https://doi.org/10.1080/09617353.2015.11691046
Published online: 11 Mar 2016.
Submit your article to this journal
Article views: 437
View related articles
View Crossmark data
Citing articles: 2 View citing articles
Full Terms & Conditions of access and use can be found at
https://www.tandfonline.com/action/journalInformation?journalCode=tsar20
5
HEART
A PROPOSED
METHOD FOR
ACHIEVING HIGH RELIABILITY IN PROCESS OPERATION
BY MEANS OF HUMAN FACTORS ENGINEERING TECHNOLOGY
J.
c.
Williams
Central Electricity Generating Board
Abstract
One of the last remaining hurdles to be overcome in the design of safe,
reliable systems is the human being. For some time now safety and
reliability engineers have recognised that they not only need to be able to
assess the likelihood and impact of human unreliability but would benefit
from a working knowledge of ho~ to apply human factors engineering
technology to optimise overall systems design.
Most safety and reliability engineers also appreciate that beyond the
immediate superficial level human engineering principles are not only
difficult to enunciate, understand or evaluate, but are occasionally costly
to apply. Often sub-systems which are manifestly well-engineered from a
human factors standpoint fail to achieve the sort of total reliability
expected when integrated with other sub-systems, to the consternation of
the process and system designer. When this happens the operator finds it
hard to believe that such a mismatch could have occurred because of a high
level failure of human engineering, and will often resort to remedies from
the management ""tool-kit"", rather than the human factors design database.
This paper attempts to minimise the likelihood of such outcomes by
identifying the relevant factors, which from the author's experience, are
considered likely to facilitate thz achievement of high man-machine
system reliability
It demystifies human factors technology by assigning
relative weights to the factors, identifying impacts and suggests a set of
human error data which should assist assessors and designers in the
achievement of high systems reliability.
1.
Introduction
In order to achieve high human reliability it is necessary to know
something about error-likely situations, what sort of error these
might lead to and the strength of the effects. Additionally it is
important to the safety and reliability engineer to have some
indication of the likely magnitude of a potential error so that
together with a knowledge of error-likely situations some remedial
measures can be taken as appropriate.
6
Until recently it has been difficult to obtain either form of
information. Whilst this paper cannot claim to furnish all the
necessary information, it does attempt to provide a simplified set
of guidelines for identifying potential major sources, types and
strengths of human error, assigning nominal probabilities of
error, identifying the impacts of some performance-shaping factors
and it suggests a battery of remedial measures which can be
invoked in order to minimise the impact of human unreliability in
process operation.
The method for conducting this analytical process is based on a
technique, developed by the author over the last 16 years, known
as HEART, the Human Error Assessment and Reduction Technique. The
technique is based p~rely on the author's-experienc; in human
factors engineering and assessment. It is not exhaustive, nor is
it validated. As an initial guide to the safety and reliability
engineer, however, it will probably be more helpful than no
guidance at all. This paper draws heavily on some notes first
prepared for the Materials Working Group of the UKAEA PWR Research
Co-ordination Committee (Williams, 1984).
2.
The Simplified HEART
2.1
Identification of sources of human unreliability
Safety and reliability engineers are concerned with gross changes
in probability of failure within systems e.g. factors of 10, the
proverbial "order of magnitude". The portion of HEART which is
likely to be of most interest to safety and reliability engineers
is concerned with those factors, therefore, which are likely to
produce probability of failure modification in excess of a factor
of 3, and which could possibly threaten system safety or
reliability. In addition to identifying the strengihs of these
factors a great many others will be mentioned, which whilst
failing to satisfy the factor of 3 criterion are important to the
assessor, precisely for this latter reason. Quite a number of
these latter performance-shaping factors are frequently mentioned
in the literature in the context of large changes, but as will be
seen from the assessor's point-of-view their effects are
relatively small and may be ignored for most practical safety and
reliability purposes.
A simplified qualitative guideline for identifying the likely
sources, classes and strengths of human error would look something
like this:-
7
Principal
Classes of
of Error
Strength
of Effect
Impaired System
Knowledge
Substitution
Omission
Insertion
Very great, especially
if a model or stereotype
is violated
Response Time
Shortage
Omission
Substitution
Timing
Insertion
Great, if system is
unforgiving
Poor or ambiguous
System feedback
Omission
Transposition
Substitution
Strong
Significance
judgement required
of operator
Omission
Substitution
Multiple
Mixed
Measurable
Level of alertness
resulting from
duties, ill-health
or environment
Omission
Substitution
Transposition
Comparatively
small
Source of Human
Unreliability
2.2
Relative Strengths of Error-producing Conditions
In general the last two sources of human unreliability do not
matter greatly from a safety snd reliability perspective because
the maximum change in unreliability which might occur will rarely
exceed a factor of 2.
This then is the first part of the HEART screening process. The
assessor should decide whether the possible sources of
unreliability are of the first three types and he can do this by
asking whether any of the following error-producing conditions
exist, assessing how much of any given condition might affect the
operator and determining from the strength of effect shown below
for each factor how much unreliability might change going from
good conditions to bad (excluding extremes such as life or healththreatening conditions).
8
Condition
Maximum predicted nominal
amount by which unreliability
might change going from good
conditions to bad
Unfamiliarity with a situation
which is potentially important
but which only occurs
infrequently or which is novel
X 17
2.
A shortage of time available
for error detection and correction
X 11
3.
A low signal-noise ratio
X
10
4.
A means of suppressing or
over-riding information or features
which is too easily accessible
X
9
s.
No means of conveying spatial
and functional information to
operators in a form which they
can readily assimilate
X
8
6.
A mismatch between an operator's
model of the world and that imagined
by a designer
X
8
7.
No obvious means of reversing
an unintended action
X
8
8.
A channel capacity overload,
particularly one caused by
simultaneous presentation
of non-redundant information
X
6
9.
A need to unlearn a technique
and apply one which requires the
application of sn opposing
philosophy
X
6
10.
The need to transfer specific
knowledge from task to task without
loss
X
5.5
11.
Ambiguity in the required
performance standards
X
5
12.
A mismatch between perceived
and real risk
X
4
13.
Poor, ambiguous or ill-matched
system feedback
X
4
9
Error-producing Condition
Maximum predicted nominal
amount by which unreliability
might change going from good
conditions to bad
X
4
Operator inexperience (e.g. a newlyqualified tradesman, but not an "expert")
X
3
16.
An impoverished quality of information
conveyed by procedures and person/person
interaction
X
3
17.
Little or no independent
checking or testing of output
X
3
14.
No
of
of
be
15.
clear direct and timely confirmation
an intended action from the portion
the system over which control is to
exerted
The following error-producing conditions are presented simply
because they are frequently mentioned in the human factors
literature as being of some importance in human reliability
assessment. To a human factors engineer, who is sometimes
concerned about performance differences of as little as 3%, all
these factors are important, but to safety and reliability
engineers who are usually concerned with differences of more than
300%, they are not very significant. The factors are identified
so that safety and reliability engineers can decide whether or not
to take account of them after the initial screening.
18.
A conflict between immediate
and long-term objectives
X
2.5
19.
No diversity of information
input for veracity checks
X
2.5
20.
A mismatch between the educational
achievement level of an individual
and the requirements of the task
X
2
21.
An incentive to use other
more dangerous procedures
X
2
22.
Little opportunity to exercise
mind and body outside the illllllediate
confines of a job
X
1.8
23.
Unreliable instrumentation (enough
that i t is noticed)
X
1.6
24.
A need for absolute judgements
which are beyond the capabilities or
experience of an operator
X
1.6
10
Error-producing Condition
2.3
Maximum predicted nominal
amount by which unreliability
might change going from good
conditions to bad
25.
Unclear allocation of function and
responsibility
X
1.6
26.
No obvious way to keep track
of progress during an activity
X
1.4
27.
A danger that finite physical
capabilities will be exceeded
X
1.4
28.
Little or no intrinsic meaning in a
task
X
1.4
29.
High-level emotional stress
X
1.3
30.
Evidence of ill-health amongst
operatives, especially fever
X
1.2
31.
Low workforce morale
X
1.2
32.
Inconsistency of meaning of displays
and procedures
X
1.2
33.
A poor or hostile environment (below
75% of health or life-threatening
severity)
X
1.15
34.
Prolonged inactivity or highly
repetitious cycling of low
mental workload tasks
X
1.1
X
1.05
35.
Disruption of normal
work-sleep cycles
X
1.1
36.
Task pacing caused by the
intervention of others
X
1.06
37.
Additional team members over and
above those necessary to perform task
normally and satisfactorily
x
1.03
38.
Age of personnel performing
perceptual tasks
X
1.02
for 1st halfhour
for each hour
thereafter
per
additional
man
Orders of Magnitude of Failure
Having determined whether the three principal sources of human
unreliability, imparted system knowledge, response time pressure
or poor or ambiguous feedback are likely to affect the achievement
11
of high reliability, the assessor is likely to want to determine
the extent of any shortfall.
He can do this by using the second part of the HEART method.
First he should decide what the likely nominal range of human
unreliability might be in relation to the types of task he is
considering. This can be achieved by consulting the following
list to see if any of the task descriptions match those he is
assessing.
Generic Task
(A)
Proposed nominal
Human Unreliability
Totally unfamiliar,
performed at speed
with no real idea of
likely consequences
5th - 95th
Percentile Bounds
0.55
0.35 - 0.97
(B)
Shift or restore system
to a new or original
state on a single
attempt without supervision
or procedures
0.26
0.14 - 0.42
(C)
Complex task requiring
high level of comprehension
and skill
0.16
0.12 - 0.28
(D)
Fairly simple task
performed rapidly or
given scant attention
0.09
0.06 - 0.13
(E)
Routine, highly-practised,
rapid task involving
relatively low level of
skill
0.02
0.007 - 0.045
(F)
Restore or shift a system
to original or new state
following procedures, with
some checking
0.007
0.0008 - 0.0035
(G)
Completely familiar, welldesigned, highly practised,
routine task occurring
several times per hour,
performed to highest possible
standards by highly-motivated,
highly-trained and experienced
person, totally aware of
implications of failure, with
time to correct potential
error, but without the benefit
of significant job aids
0.0004
0.00008 - 0.009
12
Proposed nominal
Human unreliability
(H)
Respond correctly to
system command even when
there is an augmented or
automated supervisory
system providing accurate
interpretation of system
state
0.00002
5th - 95th
Percentile Bound
0.000006 - 0.0009
If none of these task descriptions fit the type of task that
the safety and reliability engineer is considering he may take
the following values as reference points:(M)
Miscellaneous task
for which no description
can be found
0.03
0.008 - 0.11
Now he may proceed to employ the second part of the HEART
principles.
Simply by using the proposed nominal human unreliability from one
of the task descriptions above the safety and reliability engineer
can now assess not only the basic unmodified likelihood of task
failure, but he can examine what might happen to this assessed
value if some or all of the error-producing conditions are present
to any extent.
To calculate the effect of the error-producing conditions all the
assessor need do is estimate using his own judgement, that of a
group or that of a human factors engineer, what proportion of any
given error-producing condition might exist and multiply the basic
task unreliability by the appropriate proportions of the errorproducing conditions.
A safety and reliability engineer wishes to assess the nominal
likelihood of an operative's failing to isolate a plant bypass
route following strict procedures, but when it necessitates a
fairly inexperienced operator's applying an opposite technique to
that which he normally uses to carry out isolations and involves a
piece of plant, the inherent major hazards of which he is only
dimly aware. We shall assume that the man could be in the seventh
hour of the shift, that there is talk of the plant's imminent
closure, that his work may be checked and that the local
management of the company is desperately trying to keep the plant
operational despite the real need for maintenance because of its
fear that partial shutdown could quickly lead to total permanent
shutdown.
13
Using a simplified HEART the safety and reliability engineer's
assessment would look something like this:Nominal Human Unreliability
Type of Task
o.oo7
F
Error-Producing Conditions
lst screening
Factor
Total HEART
Affect
Inexperience
Opposite technique
Risk misperception
Assessed Proportion
of Affect
3
6
0.4
X
X
4
0.8
X
(3-l)
(6-1)
(4-1)
1.0
Assessed
Affect
X
X
X
0.4 + 1 = 1.8
1.0 + 1 a 6.0
0.8 + 1 • 3.4
Assessed nominal likelihood of failure
0.007
X
1.8
X
6
X
3.4 • 0.26
If our engineer took account of the other factors his assessment
could add:Factor
Total HEART
Affect
Conflict of
objectives
Low morale
Assessed Proportion
of Affect
Assessed
Affect
= 2.2
X
2.5
0.8
(2.5-1)
X
0.8 +
X
1.2
0.6
(1.2-1)
X
0.6 + 1 • 1.02
Assessed nominal likelihood of failure
0.007
X
1.8
X
6.0
X
3.4
X
2.2
X
1.02
0.58
Time-on-shift effects would be ignored as there's no indication of
monotony.
Similar calculations may be performed if desired to calculate the
predicted 5th and 95th percentile bounds.
(As a total probability of failure can never exceed 1.00, if the
multiplicstion of factors takes the value above 1.00, the probability
of failure has to be assumed to be 1.00 and no more).
14
The relative contribution made by each of the error-producing
conditions to the amount of unreliability modification is as
follows:Technique unlearning
Misperception of risk
Conflict of objectives
Inexperience
Low morale
42
24
15
12
% contribution made
to unreliability
modification
7
and the conclusions that may be drawn from this assessment (if
correct) should be fairly obvious.
It is totally inappropriate to require the application of an
opposing philosophy to effect this isolation, it is crucial to
alert operators to the hazards associated with this part of the
plant and it is somewhat unreasonable to impose a doctrine of high
availability on an inexperienced worker and his colleagues who may
be placed in a difficult position when trying to perform their
duties.
It should also be apparent that the initial screening was
sufficient to tell us the dominant factors to which the safety and
reliability engineer should attend.
Should we wish to reduce the predicted unreliability without
changing any of the error-producing conditions we would have to
introduce an automated or augmented supervisory system to provide
accurate interpretations of the plant state. The HEART method
predicts that the introduction of such a design modification
(without modification of the assessed error-producing conditions)
should reduce the nominal unreliability to 0.0016.
As the assessed nominal unreliability for an unmodified F-type of
generic task was predicted to be 0.007 we now have· a means of
comparing the cost-benefits associated with (say) modifying this
particular plant isolation system to make it work the same way as
all the other portions of plant, alerting operators to the hazards
associated with this portion of plant etc. versus introducing an
automated or augmented supervisory system. It would seem
reasonable to suppose that these activities would at least be
partially successful and allow us to achieve much higher human
reliability than might otherwise be the case.
If we wished we could explore the possibilities using the same
assessment technique.
2.5
Remedial Measures
It will have become apparent that application of the Human Error
Assessment part of the HEART method suggests possible-errorReduction Techniques that could be employed, either to combat the
predicted effects of the error-producing conditions, or else to
15
minimise the likelihood of human error occurring in a general
sense. Additionally it will be clear that the first four types of
generic task scenario may not be acceptable when high reliability
is required during process operation, so any measures that can be
employed to suppress and control these error-producing tasks would
perhaps be worth exploring.
Some measures that can be taken both to combat error-producing
conditions and tasks are described below:-
Error-Producing Condition
1.
Unfamiliarity
(x 17)
Remedial Measure
Train operators to be aware of
infrequently-occurring conditions,
simulate such situations, and teach
an understanding of the
consequences.
2.
Time shortage
(x 11)
Management must be aware that
shortage of time is likely to
impair the reliability of
decisions, both their own and their
staff's - and try to ensure that
sensitive decisions are not taken
against the clock.
3.
Low S/N ratio
Strenuous efforts must be made to
ensure that such ratios do not
fall to unreasonably low levels
(x 10)
4.
Features over-ride
allowed
(x 9)
S.
Spatial and functional
incompatibility
(x 8)
If the consequence of placing a
system in an inappropriate state is
potentially damaging, suitable
inter-locking and inhibition must
be provided, together with any
suitable time-outs to return
features to their appropriate
quiescent state
Such incompatibilities should not
occur - sufficient is now known
about human engineering for
population stereotypes that the
problem need not arise to any
extent - where doubt exists advice
should be obtained from trained
Ergonomists, who will either know
exactly how to arrange a design for
spatial or functional
compatibility, or how to run an
appropriate experiment to find out
what is required.
16
Error-Producing Condition
Remedial Measure
6.
Model mismatch
(x 8 )
Designers of systems and equipment
aren't always right - operators
sometimes not only often have
better ideas but possess views
about how a system should function
which are contrary to those of
system designers - under pressure,
particularly, operators will revert
to their own perceptions of how a
system should function, often with
undesirable consequences - to
protect against such mismatches
systems designers must try to find
out what their users' expectations
are, and then design these
characteristics into the system,
omitting their own prejudices, as
they do so.
7.
Irreversibility
Obvious means should be provided to
ensure that errors can be reversed
easily, for preference by means of
reversing the actions which created
the error in the first place.
(x 8)
8.
Channel overload
(x 6)
9.
Technique unlearning
(x 6)
It should never be necessary to
monitor more than one information
channel at any one time - single
events should not occur at more
than three per second.
The greatest possible care should
be exercised when new techniques
are being considered to achieve the
same outcome - they should not
involve adoption of opposing
philosophies.
10. Knowledge transfer
(x 5.5)
Reliance should not be placed on
operators' transferring their
previous knowledge without loss of
precision and meaning - if such
perfect transfer is required
suitable job aids must be made
available for reference.
11. Performance ambiguity
(x 5)
The required performance standards
must be tested for
comprehensibility on the user
population to ensure that there is
no ambiguity.
17
Error-Producing Condition
12. Misperception of risk
(x 4)
13. Poor feedback
(x
4)
14. Delayed/incomplete
feedback
(x 4)
15. Inexperience
(x 3)
16. Impoverished information
(x 3)
17.
Inadequate checking
(x 3)
18. Objectives conflict
(x 2.5)
Remedial Measure
It must not be assumed that a
user's perception of risk is the
same as the actual level - if
necessary a check should be made to
ascertain where any mismatch might
exist and what its extent is.
A task analysis will show the
points at which feedback must be
available to operators Ergonomists can advise on the best
form of feedback if doubts should
arise - what one is looking for is
complete "system transparency"
System response times should never
exceed four seconds and there must
always be sufficient information to
enable operators to step
confidently on to the next part of
a task - if doubt exists the
feedback is incomplete.
Personnel criteria should contain
specified experience parameters
thought relevant to the task chances must not be taken for the
sake of expediency.
Procedures should be humanengineered and tested for
operability - it should be assumed
that when personnel are required to
communicate with each other that
very considerable information loss
will occur - procedures must not
rely on accurate verbal transmission
of information for success.
When high reliability is paramount,
independent checks on accuracy
should be made, by people and
systems that do not have any vested
interest in the success or failure
of an individual - blame should not
attach to any inadequacies found at
this level.
Objectives should be tested by
management for mutual compatibility,
and where potential conflicts are
identified these should either be
18
Error-Producing Condition
18. Objectives conflict
(Continued •••••• )
(x
2.5)
19. No diversity
(x
2.5)
20. Educational Mismatch
(x 2)
21. Dangerous incentives
(x 2)
22. Lack of exercise
(x 1.8)
Remedial Measure
resolved to make them harmonious or
made prominent so that a
comprehensive management control
programme can be created to
reconcile such conflicts as they
arise, in a rational fashion.
It should not be assumed that
operators will rely totally on a
single information source for
confirmation of accuracy, and
enquiries should be made to
ascertain what additional sources
are referred to, so that these are
not denied operators, and, if
possible, are enhanced.
The job profile should identify any
potential mismatch of recruits
against requirements - educational
standards should be made explicit;
there should be no ambiguity.
It is intuitively obvious that
people work for rewards of various
natures - if the reward for doing
something quickly is greater than
the reward for doing it accurately,
or the reward for omitting an action
is greater than the reward for
performing it we should not be
surprised if that is, in the main,
what happens - the reward system
must be evaluated carefully,
therefore, to ensure that the
desired behaviour is emitted,
rather than that which might be
construed as being appropriate
simply because facets of the task
are seen to conform to a partial
criterion - if in doubt, seek
advice from Management Scientists
and/or Psychologists.
Frequent rest breaks should be
designed into the job, and the
system made tolerant to personnel
taking breaks as the need arises tuition should be given in
techniques for maintaining high
levels of arousal, such as postural
19
Error-Producing Condition
Remedial Measure
22. Lack of exercise
(Continued •••••• )
(x 2)
change, personal ventilation and
recognition of fatigue symptoms encouragement should be given to
engage in appropriate mild forms of
physical exercise and relaxation
and stress control - on-the-job
refresher training should be given
and frequent exercises to maintain
and enhance levels of competence
and awareness of technical progress
innovation given.
23. Unreliable instruments
Regrettably it is a fact that when
instrumentation is found to be
unreliable operators will cease to
trust its indications to the extent
of ignoring valid information and
preferring to believe their own
interpretations, despite
overwhelming evidence to the
contrary - if instrumentation is
thought likely to be unreliable it
should be withdrawn from service,
and more reliable instrumentation
substituted - no doubts should
exist about its suitability.
(x 1.6)
24. Absolute judgements
required
(x 1.6)
Operators must not be placed in the
position of having to make
judgements about the meaning of
data which are outside their span
of apprehension or experience - a
task analysis will reveal when such
conditions are likely to arise, and
management must plan for such
contingencies, by recognising the
circumstances and taking full
responsibility for actions which
might be taken on their behalf "brain-storming" and problemsolving workshops are helpful to
identify some of the most bizarre
situations in which staff and
management can find themselves- it
is likely that discussion of these
'grey areas' of organisational
behaviour will reinforce mutual
respect, and anticipate future
conflict and/or issues of
culpability at a time of zero
threat.
20
Error-Producing Condition
25. Unclear allocation of
Function
(x 1.6)
26. Progress tracking lack
(x 1.4)
27. Physical capabilities
(x 1.4)
28. Low Meaning
(x 1.4)
Remedial Measure
As with the area above, doubt must
not exist about responsibilities whilst they can, and should, be
stated on paper, joint preparation
of a functional specification will
remove doubts and anxieties, and
lead to the development of healthy
attitudes towards the system design
concepts - Organisational
Development Specialists and/or
Behavioural Scientists should be
involved in facilitating the
preparation of a satisfactory
working protocol.
Various job aids must be supplied
in order to ensure that operators
do not get out of step with the
task in hand - these can range from
checklists through mimics to
electronic monitoring of progress
against targets - if such aids are
introduced they must be piloted to
ensure that they are compatible
with user needs and that there is
an incentive to use them Ergonomists can advise on these job
design aspects.
It should be self-evident that
tasks must not exceed the
operators' capabilities - reference
to Human Factors Standards will
ensure that these capabilities are
not exceeded.
Meaning can be built into a job by
preparing job descriptions with
the staff concerned, showing them
the significance of their
contribution to corporate
objectives, designing variety into
their duties by arranging for job
features such as task rotation to
enhance system awareness, and
holding periodic reviews of working
practices to ensure that symptoms
of alienation are not manifesting
themselves - Behavioural Scientists
can advise on suitable
precautions.
21
Error-Producing Condition
Remedial Measure
29. Emotional stress
(x 1.3)
Management and medical staff must
be vigilant to recognise the onset
of emotional problems which can
manifest themselves via symptoms
such as excessive absence,
persistent lateness, obsessive
behaviour, lack of co-operation and
exceptional fatigue - personal
stress control training programmes
could be considered, and
potentially stressful decisionmaking circumstances identified so
that the conditions can be modified
to limit occurrence of extreme
generalised stress.
30. Ill-health
Until it is pointed out, it is not
apparent that ill-health can have
such deleterious effects on
performance- often the effects of,
say, a cold or 'flu do not manifest
themselves until well into a shift
- by now it should be obvious that
operators and managers who are ill
should not attempt to undertake
work requiring high reliability,
and out of respect for others, for
system integrity and peace of mind
they should stay away, until
recovered - a medical awareness
programme would be helpful.
(x 1.2)
31. Low morale
(x 1.2)
Apart from the more obvious ways of
attempting to secure high morale by
way of financial reward, for
example, other methods involving
participation, trust and mutual
respect, often hold out at least as
much promise - building up morale
is a painstaking process, which
involves a little luck and great
sensitivity - employees must be
given reason to believe in their
employer and themselves - this can
be accomplished by a battery of
activities, such as joint
preparation of work plans and
objectives, maximal delegation of
authority, reward for effort and
results, provision of subsidised
fringe benefits, firmness of
22
Error-producing Condition
31. Low morale (Continued ••• )
(x 1.2)
32. Inconsistency of
Displays
(x
1.2)
33. Poor environment
(x 1.15)
34. Low loading
(x 1.1) 1st\ hr.
(x 1.05) each hour
thereafter
Remedial Measure
resolve and openness -it is not
achieved to any great extent by
appeals to workforces to stick by
management - the respect necessary
to make morale rise is earned not
enforced - a sensitive, caring
management, would be unlikely to
encounter such problems.
Even if the conventions adopted for
display layout and procedure design
are not human-engineered for ease
of use, they must be consistent
within themselves e.g. if a display
is showing an increasing value even
though in an analogue sense the
portion shown is decreasing, this
convention must be adhered to
throughout - even though such a
principle is "wrong" (for
preference such an approach would
not be encouraged, of course)
It should be self-evident that a
poor environment is likely to impair
performance - by and large this
should not occur nowadays with the
introduction of legislation to
control environments - to minimise
any deleterious effects Work
Physiologists, Ergonomists and/or
Architects should be consulted for
details of appropriate parameters.
Prolonged inactivity or highly
repetitious cycling of low mental
workload tasks must be avoided generally when signal frequency
falls below two per minute or
involves little or no variability,
vigilance performance will degrade
- to combat such effects the
introduction of artificial signals
has been found to be helpful, and
job enrichment (with the
in(4oduction of different, more
varied tasks) has been found to
minimise boredom, and better hold
attention - rather than combat
these effects, it is better to
ensure that such conditions do not
arise in the first place i.e.
23
Error-producing Condition
34. Low loading (Continued •• )
(x 1.1 ) 1st
~hr.
(x 1.05) each hour
thereafter
35. Sleep cycle disruption
(x 1.1)
Remedial Measure
observation tasks demanding high
human reliability should never
require sessions of longer than one
hour's concentration and tasks
involving very low signal frequency
should not be designed - if
possible such tasks should be
automated.
Only extreme sleep deprivation will
cause performance degradation - our
major interest, therefore, is in
keeping small amounts of
deprivation to a minimum - this can
be achieved by keeping operators on
a "stable" shift system such that
there are no radical changes to
either the pattern or the time of
day over which such changes occur the frequency with which
changeovers occur should be as low
as can reasonably be achieved advice should be sought from Work
Physiologists.
36. Task Pacing
(x 1.06)
Although all work ultimately
involves some element of pacing,
the unwitting or deliberate
introduction of pacing will lead to
a slight reduction in reliability this can be avoided by checking
work systems to ensure that there
is sufficient 'buffering' such that
operators are not subject to undue
pressure and can work at their own
preferred pace - the one which best
matches their capability.
37. Supernumeraries
Where possible, limit gatherings
of staff at workplaces to those
necessary to perform tasks
satisfactorily.
(x 1.03)
38. Age
(x 1.02)
Monitor perceptual capabilities of
personnel required to perform task
demanding high acuity and
accurate information processing.
24
2.6
Suppression and control of error-producing tasks
Unfamiliar tasks
Plan the training and work programme
so that the likelihood of new,
unfamiliar tasks arising is extremely
small.
Restoration of System
State
As mentioned elsewhere, care should be
taken to ensure that there is a system
of auto time-outs, inhibitions,
defaults etc. to ensure that wherever
possible, responsibility does not rest
on the operator to shift or restore
the system to a new or original state,
without supervision - obviously, where
possible, job aids and supervision
should be available to unburden the
operator.
Complex Task
Clearly the hope is that operators
will be become sufficiently proficient
that the sort of task identified will
not prove to be a complex one,
requiring a high level of
comprehension and skill- if such a
task exists in their repertoire then
the training might be judged to have
failed - this potential problem then
should be removed by training to some
performance criterion.
Simple Task
For high reliability this is what we
need our tasks to amount to in the
minds of our operators, but we must
ensure that they are not performed in
a cavalier fashion -t'herefore whilst
training them with the hope that the
tasks will ultimately appear simple,
an integral part of the training
programme will be to alert operators
to the pitfalls of thinking that
simple tasks can be performed rapidly
in this context, or given scant
attention.
Highly-Practised
Below this level there is little that
can be achieved to control the
production of error - operators will
already be performing at the limits of
their ability - further improvements
can only be achieved by means of
constant rehearsal and training to
appreciate the "cost" of failure or
the introduction of system aiding
and/or increasing amounts of
supervision.
25
3.
Conclusions
By now it should be clear that the safety and reliability engineer
is likely to be in a position to achieve high human reliability
in process operation. Whilst the HEART method is so far
unvalidated, it does at least hold out the hope of indicating the
likely magnitude of potential errors, and identifying some of the
error-likely situations.
The simplified set of guidelines could serve as a means of
identifying potential major sources, types and strengths of human
error, whilst application of the battery of remedial measures
should give some prospect of achieving high human reliability in
process operation.
The perennial problems associated with assessing human
reliability, namely the relative absence of data and validation
have not receded, but there is some hope that data sources may be
found and that the need for validation will be acted upon in the
near future. In the meantime the safety and reliability engineer
must be content with somewhat inadequate tools. However,
inadequate the HEART method may prove to be it is almost certainly
more useful than no guidance at all.
4.
Reference
Williams (1984)
"Guidelines on possible sources of Human Error" PWR
Research Co-ordination Committee, Materials Working Group,
PWR/RCC/MWG/P(84)405, 10 December 1984, UKAEA, Risley,
Warrington.
5.
Acknowledgements
The author would like to thank Miss C Joures for typing this
manuscript.
N.B.
The views expressed in this paper are not necessarily those
of the Central Electricity Generating Board.
Download