FR_Principles_CERN_110505

advertisement
PRINCIPLES OF
SYSTEM SAFETY
ENGINEERING AND MANAGEMENT
Felix Redmill
Redmill Consultancy, London
Felix.Redmill@ncl.ac.uk
RISK
(c) Felix Redmill, 2011
CERN, May '11
2
SAFETY ENGINEERING AND MANAGEMENT
It is necessary both to achieve appropriate safety and to
demonstrate that it has been achieved
• Achieve - not only in design and development, but in all
stages of a system’s life cycle
• Appropriate - to the system and the circumstances
• Demonstrate - that all that could reasonably have been
done has been done, at every stage of the life cycle
(c) Felix Redmill, 2011
CERN, May '11
3
THE U.K. LAW ON SAFETY
Health and Safety at Work Etc. Act 1974:
Safety risks imposed on others (employees and the public
at large) must be reduced ‘so far as is reasonably
practicable’ (SFAIRP)
(c) Felix Redmill, 2011
CERN, May '11
4
THE HSE’S ALARP PRINCIPLE
Increasing
risk
Unacceptable Region
(Risk cannot be justified except in
extraordinary circumstances)
Limit of tolerability threshold
ALARP or Tolerability Region
(Risk is tolerable only if its reduction is
impracticable or if the cost of reduction is
grossly disproportinate to the improvement
gained)
Broadly acceptable threshold
Broadly Acceptable Region
(Risk is tolerable without reduction. But it is
necessary to maintain assurance that it
remains at this level)
(c) Felix Redmill, 2011
CERN, May '11
5
THE HSE’S ALARP PRINCIPLE
(c) Felix Redmill, 2011
CERN, May '11
6
CALIBRATION OF THE ALARP MODEL
Recommended for Nuclear Industry
• Intolerability threshold:
– 1/10000 per year (for the public)
– 1/1000 per year (for employees)
• Broadly acceptable threshold:
– 1/1000000 per year (for everyone)
(c) Felix Redmill, 2011
CERN, May '11
7
A VERY SIMPLE SYSTEM
• Chemicals A and B are mixed in the tank to form product P
• P opens & closes input and output valves
• If emergency signal arrives, then cease operation
Emergency
signal
B
P
A
P Controller
(c) Felix Redmill, 2011
CERN, May '11
8
QUESTIONS
• How could the accident have been avoided?
– Better algorithm
• How could the software designer have known that a better
algorithm was required?
– Domain knowledge
• But we can’t be sure that such a fault won’t be made, so
how can we find and correct such faults?
– Risk analysis techniques
(c) Felix Redmill, 2011
CERN, May '11
9
A SIMPLE THOUGHT EXPERIMENT
• Draw an infusion pump
– Note its mechanical parts
• Think about designing and developing software to control
its operation
• Safety consideration: delivery of either too much or too
little of a drug could be fatal
• How would you guarantee that it would kill no more than 1
in 10,000 patients per year?
– What level of confidence - on a percentage scale - do
you have in your estimate?
(c) Felix Redmill, 2011
CERN, May '11
10
CONFIDENCE IN SAFETY IN ADVANCE
• What is safety?
• How can it be measured?
• What can give confidence that safety is high?
• Need also to demonstrate safety in advance
• Therefore, need to find hazards in advance
(c) Felix Redmill, 2011
CERN, May '11
11
NEW RISKS OF NEW TECHNOLOGY STRASBOURG AIR CRASH
• Mode: climbs and descents in degrees to horizontal
– 3.3 descent represented as ‘minus 3.3’
• Mode: climbs and descents in units of 100 feet
– 3300 feet/minute descent represented as ‘minus 33’
• Plane was descending at 3300 feet/minute
• Needed to descend at angle of 3.3 to horizontal
• To interpret correctly, pilot needed to know the mode
• ‘Mode error’ in the system
• Human error? If so, which human made it?
(c) Felix Redmill, 2011
CERN, May '11
12
CONCERN WITH TECHNOLOGICAL RISKS
• We are concerned no longer exclusively with making nature
useful, or with releasing mankind from traditional
constraints, but also and essentially with problems resulting
from techno-economic development itself [Beck]
• Risks induce often irreversible harm; not restricted to their
places of origin but threaten all forms of life in all parts of the
planet; in some cases could affect those not alive at the time
or place of an accident [Beck]
• People who build, design, plan, execute, sell and maintain
complex systems do not know how they may work [Coates]
(c) Felix Redmill, 2011
CERN, May '11
13
RISK - AN IMPORTANT SUBJECT
• Risk is a subject of research in many fields,
e.g. Psychology, Sociology, Anthropology, Engineering
• It is a critical element in many fields
e.g. Geography (climate change), Agriculture, Sport,
Leisure activities, Transport, Government policy
• It influences government policy (e.g. on bird flu) and
local government decisions (e.g. closure of children’s
playgrounds)
• An influential Sociological theory is that it is the most
influential factor in modern society
(we are in ‘the risk society’)
• Every activity carries risk
• All decisions are ‘risky’
(c) Felix Redmill, 2011
CERN, May '11
14
FUNCTIONAL SAFETY:
ACHIEVING UTILITY AND CREATING RISK
We are concerned with safety that depends on the correct
functioning of equipment
Control
system
(c) Felix Redmill, 2011
Equipment
under control
CERN, May '11
Utility
(plus risks)
15
FUNCTIONAL SAFETY 2
Functional safety depends on hardware, software,
humans, data, and interactions between all of these
Control
system
Equipment
under control
Utility
(plus risks)
Humans (design, operation,
maintenance, etc.)
(human factors)
Environment (management, culture, etc.)
(c) Felix Redmill, 2011
CERN, May '11
16
ORGANISATIONS ARE ALSO COMPLEX
(Vincristine example)
• Child received injection for leukaemia into spine instead of
into vein
• Root cause analysis showed 40 points at which the
accident could have been averted
• Complexity of modern organisational systems
• Need to identify risks in advance
– But no standard can do this for us
• Requires a risk-based approach
(c) Felix Redmill, 2011
CERN, May '11
17
SAFETY - A DEFINITION
• Safety: freedom from unacceptable risk (IEC)
• Safety is not directly measurable
– But it may be addressed via risk
(c) Felix Redmill, 2011
CERN, May '11
18
VOCABULARY
• In common usage, “Risk” is used to imply:
– Likelihood (e.g. there’s a high risk of infection)
– Consequence (e.g. infection carries a high risk)
– A combination of the two
– Something, perhaps unspecified, to be avoided (e.g.
going out into the night is risky)
(c) Felix Redmill, 2011
CERN, May '11
19
RISK - DEFINITIONS
• Risk: A combination of the probability of occurrence and the
severity of its consequences if the event did occur (IEC)
• Tolerable risk: a willingness to live with a risk, so as to secure
certain benefits, in the confidence that the risk is one that is
worth taking and that it is being properly controlled (HSE)
(c) Felix Redmill, 2011
CERN, May '11
20
SAFETY AND RISK
• Safety is only measurable in retrospect
• Safety is gauged by trying to understand risk
• Risk, being of the future, is estimable but not measurable
• The higher the risk, the lower the confidence in safety
• We increase safety by reducing risk
• Two components of risk are significant:
– Likelihood of occurrence
– Magnitude of outcome
(c) Felix Redmill, 2011
CERN, May '11
21
SOME PRINCIPLES
• Absolute safety (zero risk) cannot be achieved
• ‘Doing it well’ does not guarantee safety
– Correct functionality  safety
– We must address safety as well as functionality
• Reliability is not a guarantee of safety
• We require confidence of safety in advance, not
retrospectively
• We must not only achieve safety but also demonstrate it
(c) Felix Redmill, 2011
CERN, May '11
22
A RISK-BASED APPROACH
• If safety is addressed via risk
– We must base our safety-management actions on risk
• We must understand the risks in order to manage safety
(c) Felix Redmill, 2011
CERN, May '11
23
SAFETY AS A WAY OF THINKING
• If risk is too high, it must be reduced
• But how do we know how high it is?
– Carry out risk analysis
• But even if the risk is high, does it need to be reduced?
– We must understand what risk is tolerable in the
circumstances (in the UK, apply the ALARP Principle)
• Achieving safety demands a combination of engineering
and management approaches
(c) Felix Redmill, 2011
CERN, May '11
24
RISK - TWO COMPONENTS
• Two components:
– Probability (likelihood) of occurrence
– Consequence (magnitude of outcome)
• R = f(P.C) or f(L.C)
(c) Felix Redmill, 2011
CERN, May '11
25
A SIMPLE CALCULATION
Probability of a 100-year flood = 0.01/year
Expected damage = £50M
R (financial (Expected Value)
= 0.01 x £50,000,000 = £500,000/year
(c) Felix Redmill, 2011
CERN, May '11
26
CONFIDENCE IN RISK VALUES
 Accuracy of the result depends on the reliability of information
 Reliability of information depends on the pedigree of its source
• What are the sources of the probability and consequence
values?
• What are their pedigrees?
• What confidence do we have in the risk values?
• Why do we need confidence?
– Because we derive risk values in order to inform decisions
(c) Felix Redmill, 2011
CERN, May '11
27
CONFIDENCE IN RISK CALCULATIONS
• Where did the information come from?
– What trust do we have in the source?
– When and by whom was it collected?
– Was it ever valid? If so, is it still valid?
• What assumptions have we made?
– How valid are they?
– Are we aware of them?
(c) Felix Redmill, 2011
CERN, May '11
28
COT DEATH CASE
• A study concluded that the risk of a mature, non-smoking,
affluent couple suffering a cot death is 8543 to 1
• Prof. Meadows deduced that Pr. of two cot deaths in the
same family = 8543 x 8543 = 73 million to one
• Three cot deaths in her family resulted in Mrs Clark being
convicted of infanticide
(c) Felix Redmill, 2011
CERN, May '11
29
DEPENDENCE AND NOT RANDOMNESS
• But deaths in the same family are not independent events
(probability theory assumes independence)
• One death in the family rendered Mrs Clark “in the highest-
risk category” for another
– Probability of a second death is considerably greater
than 8543 to 1
(c) Felix Redmill, 2011
CERN, May '11
30
COMMON-MODE FAILURES
• Identifying common-mode failures is a crucial part of
traditional risk analysis
(c) Felix Redmill, 2011
CERN, May '11
31
ASSUMPTIONS
• We never have full knowledge
• We fill the gaps with assumptions
• Assumptions carry uncertainty, and therefore risk
• Recognise your assumptions (if possible)
• If you admit to them (document them)
– Readers will recognise uncertainty
– Other persons may provide closer approximations
(c) Felix Redmill, 2011
CERN, May '11
32
CONTROL OF RISK
• Risk is eliminated if either Pr or C is reduced to zero
• And risk is reduced by reduction of Pr or C or both
• In many cases we have no control over C, but we may be
able to estimate it
• It may be difficult or impossible to derive Probability, but
we may be able to reduce it (e.g. by software testing &
fixing)
(c) Felix Redmill, 2011
CERN, May '11
33
WHY TAKE RISKS?
• Progress demands risk
– e.g. exploration, bridge design
– e.g. drug development (TeGenero’s TGN1412)
• Technology provides utility - at the cost of risk
• Suppliers want cheap designs
– e.g. Ronan Point (need to foresee hazards)
• To save money
– e.g. Cutting corners on a project
• To save face
– E.g. “I’d rather die than make a fool of myself”
• We can’t avoid taking risks in every decision and action
(c) Felix Redmill, 2011
CERN, May '11
34
RISK VS. RISK
• Decision is often not between risk and no risk
– But between one risk and another
– Decisions may create risks (unintended consequences)
• Surgery, medication, or keep the illness
• Consequences of being late vs. risks of driving fast
• Solar haze vs. global warming
• Software maintenance can introduce a new defect
– Change creates a new product; test results obsolete
• Recognise both risks
• Assess both
• Carry out impact analysis to identify new hazards
(c) Felix Redmill, 2011
CERN, May '11
35
SOME NOTES ON RISK
• A single-system risk may be tiny, but the overall risk of many
systems may be high (individual vs. societal risk)
• Risk per usage may be remote, but daily use may accrue a
high risk (one-shot vs. lifetime risk)
• Beware of focusing on a single risk; a system is likely to carry
numerous risks
• Confidence in probabilities attributed to rare events must be
low
• For random events with histories (e.g. electromechanical
equipment failure) frequencies may be deduced
– Given similar circumstances, they may be predictive
• For systematic events (such as software failure) history is not
an accurate predictor of the future
(c) Felix Redmill, 2011
CERN, May '11
36
RISK AND UNCERTAINTY
• Risk is of the future - there is always uncertainty
• Risk may be estimated but not measured
• Risk is open to perception
• Identifying and analysing risks requires information
• The quality of information depends on the source
• 'Facts' are open to different interpretations
• Risk cannot be eliminated if goals are to be achieved
• Risk management activities may introduce new risks
• In the absence of certainty, we need to increase confidence
• Reducing uncertainty depends on gathering information
(c) Felix Redmill, 2011
CERN, May '11
37
RISK MANAGEMENT STRATEGIES
•
•
•
•
•
Avoid
Eliminate
Reduce
Minimise - within defined constraints
Transfer or share
- Financial risks may be insured
- Technical risks may be transferred to experts, maintainers
• Hedge
- Make a second investment, which is likely to succeed if the
first fails
• Accept
- Must do this in the end, when risks are deemed tolerable
- Need contingency plans to reduce consequences
(c) Felix Redmill, 2011
CERN, May '11
38
RISK IS OPEN TO PERCEPTION
•
•
•
•
•
•
•
•
•
•
Voluntary or involuntary
Control in hands of self or another
Statistical or personal
Level of knowledge, uncertainty
Level of dread or fear evoked
Short-term or long-term view
Severity of outcome
Value of the prize
Level of excitement
Status quo bias
 Perception determines where we look for risks and what
we identify as risks
(c) Felix Redmill, 2011
CERN, May '11
39
PROGRAMME FOR COMBATTING DISEASE
(1)
• The country is preparing for the outbreak of a disease
which is expected to kill 600 people. Two alternative
programmes to combat it have been proposed, and the
experts have deduced that the precise estimates of the
outcomes are:
• If programme A is adopted, 200 people will be saved
• If programme B is adopted, there is a one third probability
that 600 people will be saved and a two thirds probability
that nobody will be saved
(c) Felix Redmill, 2011
CERN, May '11
40
PROGRAMME FOR COMBATTING DISEASE
(2)
• The country is preparing for the outbreak of a disease
which is expected to kill 600 people. Two alternative
programmes to combat it have been proposed, and the
experts have deduced that the precise estimates of the
outcomes are:
• If programme C is adopted, 400 people will die
• If programme D is adopted, there is a one third probability
that nobody will die and a two thirds probability that 600
people will die
(c) Felix Redmill, 2011
CERN, May '11
41
UNINTENDED CONSEQUENCES
• Actions always likely to have some unintended results
– Downsizing but loose the wrong staff
 e.g. University redundancies in UK
– Staff come to rely entirely on a support system
– Building high in mountain causes landslide
– Abolishing DDT increased malaria
– Store less chemical at plant, more trips by road
• Unintended, but not necessarily unforeseeable
• Foreseeable, but not necessarily obvious
• Carry out hazard and risk analysis
– Better to foresee them than encounter them later
(c) Felix Redmill, 2011
CERN, May '11
42
RISK COMMUNICATION
• The risk analyst is usually not the decision maker
– Risk information must be transferred (liver biopsy)
• Usually only results are communicated
– e.g. There’s an 80% chance of success
• But are they correct (Bristol Royal Infirmary)?
– Overconfidence bias
• Managers rely heavily on risk information, collected,
analysed, packaged, by other staff
• But what confidence does the staff have in it
– What were the information sources, analysis methods,
and framing choices?
– What uncertainties exist?
– What assumptions were made?
(c) Felix Redmill, 2011
CERN, May '11
43
COMMUNICATION OF RISK INFORMATION
• ‘The risk is one in a million’
• One in seventeen thousand of dying in a road accident
– Age, route, time of day
• Framing
• Appropriate presentation of numbers
– 1 x 10-6
– One in a million
– One in a city the size of Birmingham
– (A launch per day for 300 years with only one failure)
(c) Felix Redmill, 2011
CERN, May '11
44
RISK IS TRICKY
• We manage risks daily, with reasonable success
• Our risk management is intuitive
– We do not recognise mishaps as the results of poor
risk management
• Risk is a familiar subject
• We do not realise what we don’t know
• We do not realise that our intuitive techniques for
managing simple situations are not adequate for complex
ones
• Risk estimating can be simple arithmetic
• The difficult part is obtaining the correct information on
which to base estimates - and decisions
(c) Felix Redmill, 2011
CERN, May '11
45
SAFETY ENGINEERING PRINCIPLES
(c) Felix Redmill, 2011
CERN, May '11
46
A SIMPLE MODEL OF SYSTEM SAFETY
Failure
or unsafe
deviation
Safe state
(in any
mode)
Accident
Danger
Disaster
Restoration
Recovery
(c) Felix Redmill, 2011
CERN, May '11
47
SAFETY ACROSS THE LIFE CYCLE
• Safety activities must extend across the life of a system
– And must be planned accordingly
• Modern safety standards call for the definition and use of a
safety life cycle
• A life-cycle model shows all the phases of a system’s life,
relative to each other
• The ‘overall safety lifecycle’ defined in safety standard IEC
61508 is the best known model
(c) Felix Redmill, 2011
CERN, May '11
48
SAFETY LIFE CYCLE MODELS
• Provide a framework for planning safety engineering and
management activities
• Provide a guide for creating an infrastructure for the
management of project and system documentation
• Remind engineers and managers at each phase that they
need to take other phases into consideration in their
planning and activities
• Facilitate the demonstration as well as the achievement of
safety
• The model in safety standard IEC 61508 is the best known
(c) Felix Redmill, 2011
CERN, May '11
49
OVERALL SAFETY LIFECYCLE
Concept
1
2
Overall scope
definition
3
Hazard and risk
analysis
4
Overall safety
requirements
5
Safety requirements
allocation
Overall planning of:
6
O&M
(c) Felix Redmill, 2011
7
Safety
validation
Realisation of:
8
Installation &
commissioning
9
Safetyrelated
E/E/PES
12
Overall installation
and commissioning
13
Overall safety
validation
14
Overall operation,
maintenance & repair
Decommissioning or
CERN, May
'11
disposal
16
10
Other tech.
safetyrelated
systems
15
11
External
risk
reduction
facilities
Overall
modification
and retrofit
50
AFTER STAGE THREE
Work done after stage 3 of the model creates additions to
the overall system that were not included in the hazard
and risk analysis of stage 3
(c) Felix Redmill, 2011
CERN, May '11
51
SAFETY LIFECYCLE SUMMARY
Understand the functional goals and design

Identify the hazards

Analyse the hazards

Determine and assess the risks posed by the hazards

Specify the risk reduction measures and their SILs

Define the required safety functions (and their SILs)

Carry out safety validation

Operate, maintain, change, decommission, dispose safely
(c) Felix Redmill, 2011
CERN, May '11
52
THE SAFETY-CASE PRINCIPLE
• The achievement of safety must be demonstrated in advance
of deployment of the system
• Demonstration is assessed by independent safety assessors
• Safety assessors will not (cannot)
– Examine a complex system without guidance
– Assume responsibility for the system’s safety
• Their assessment is guided by claims made by developers,
owners and operators of the system
• The claims must be for the adequacy of safety in defined
circumstances, considering the context and application and
the benefits of accepting any residual risks
(c) Felix Redmill, 2011
CERN, May '11
53
THE SAFETY CASE
• The purpose: Demonstration of adequate and appropriate
safety of a system, under defined conditions
– To convince ourselves of adequate safety
– The basis of independent safety assessment
– Provides later requirements for proof
(It can protect and it can incriminate)
• The means
– Make claims of what is adequately safe
o And for what it is adequately safe
– Present arguments for why it is adequately safe for the
intended purpose
– Provide evidence in support of the arguments
(c) Felix Redmill, 2011
CERN, May '11
54
GENERALIZED EXAMPLE
• Claim: System S is acceptably safe when used in Application A
• Claims are presented as structured arguments
– The use of System S in Application A was subjected to
thorough hazard identification
– All the identified hazards were analysed
– All risks associated with the hazards either were found to be
tolerable or have been reduced to tolerable levels
– Emergency plans are in place in case of unexpected
hazardous events
• The safety arguments are supported by evidence
– e.g., Evidence of all activities and results, descriptions of all
relevant documentation, and references to it
(c) Felix Redmill, 2011
CERN, May '11
55
THE CASE MUST BE STRUCTURED
• Demonstration of adequate safety of a system requires
demonstration of justified confidence in its components
• Some components may be COTS (commercial off-the-shelf)
– Don’t wait until too late to find that they cannot be justified
• Evidence is derived from different sources, and different
categories of sources
• The evidence for each claim must be structured so as to
support a logical argument for the “top-level” claim
(c) Felix Redmill, 2011
CERN, May '11
56
PRESENTATION OF SAFETY CLAIMS
• The claims must be structured so that the logic of the overall
case is demonstrated
• The principal claims may be presented in a “top-level”
document
– With references out to the sources of supporting evidence
(gathered throughout the life cycle)
• Modern software-based tools are available for the
documentation of safety cases (primarily: GSN (goal-structured
notation))
(c) Felix Redmill, 2011
CERN, May '11
57
THE NATURE OF EVIDENCE
• Evidence may need to be probabilistic
– e.g. for nuclear plants
• It may be qualitative
– e.g. for humans, software
• In applications of human control it may need to include
– Competence, training, management, guidelines
• It is collected throughout the system’s life
– Importantly during development
(c) Felix Redmill, 2011
CERN, May '11
58
EVIDENCE PLANNING - 1
• Any system-related documentation (project, operational,
maintenance) may be required as evidence
• It should be easily and quickly accessible to
– Creators of safety arguments
– Independent safety assessors
• Numbering and filing systems should be designed appropriately
• Principal safety claims should be identified in advance
– Activities may need to be planned so that appropriate
evidence is collected and stored
• The safety case should be developed throughout a development
project
– And maintained throughout a system’s life
(c) Felix Redmill, 2011
CERN, May '11
59
EVIDENCE PLANNING - 2
• The safety case structure should be designed early
• Evidential requirements should be identified early and
planned for
• Safety case development should be commenced early
• Evidence of software adequacy may be derived from
– Analysis
– Testing
– Proven-in-use
– Process
(c) Felix Redmill, 2011
CERN, May '11
60
VALIDITY OF A SAFETY CASE
Validity (and continued validity) depends on (examples only)
• Observance of design and operational constraints, e.g.:
– Use only specified components
– Don’t exceed maximum speed
• Environment, e.g.:
– Hardware: atmospheric temperature between T1 & T2C
– Software: operating system remains unchanged
• Assumptions remain valid, e.g.:
– Those underlying estimation of occurrence probability
– Routine maintenance carried out according to spec.
(c) Felix Redmill, 2011
CERN, May '11
61
HAZARD AND RISK ANALYSIS
(c) Felix Redmill, 2011
CERN, May '11
62
NEED FOR CLARITY
• Is a software bug a risk?
• Is a banana skin on the ground a risk?
• Is a bunker on a golf course a risk?
(c) Felix Redmill, 2011
CERN, May '11
63
THE CONCEPT OF A HAZARD
• A hazard is the source of risk
– The risks that may arise depend on context
• What outcomes could result from the hazard?
• What consequences might the outcomes lead to?
• Hazards form the basis of risk estimation
(c) Felix Redmill, 2011
CERN, May '11
64
GOLF-SHOT RISK
• What is the probability of hitting your golf ball into a bunker?
• What is the consequence (potential consequence) of hitting
your ball into a bunker?
• Should I “take the risk”?
– In this case: What level of risk should I take?
• What’s missing is a question about benefit!
(c) Felix Redmill, 2011
CERN, May '11
65
WHAT IS RISK ANALYSIS?
• If we take risk to be a function of Probability (Pr) &
Consequence (C)
– Then, in order to estimate a value of a Risk, we need to
derive values of Pr and C for that risk
• Values may be qualitative or quantitative
• They may need to be more or less “accurate”, depending on
importance
• Thus, risk analysis consists of:
– Identifying relevant hazards
– Collecting evidence that could contribute to the
determination of values of Pr and C for a defined risk
– Analysing that evidence
– Synthesising the resulting values of Pr and C
(c) Felix Redmill, 2011
CERN, May '11
66
PROBABILITY OF WHAT EXACTLY?
• Automobile standards focus on the control of a
vehicle (and the possible loss of it)
• Each different situation will
– Occur with a different probability
– Result in different consequences
– Carry different risks
(c) Felix Redmill, 2011
CERN, May '11
67
CHOICE OF CONSEQUENCE ESTIMATES
• Worst possible
• Worst credible
• Most likely
• “Average”
(c) Felix Redmill, 2011
CERN, May '11
68
DETERMINATION OF RISK VALUES
Threat to
security
Safety hazard
Causal
analysis
Threat of
damage
Potential for
unreliability
Risk
Consequence
analysis
Potential for
unavailability
(c) Felix Redmill, 2011
CERN, May '11
69
PRELIMINARY REQUIREMENTS
• Knowledge of the subject of risk
• Understanding of the current situation and context
• Knowledge of the purpose of the intended analysis
– The questions (e.g. tolerability) that it must answer
 Such knowledge and understanding are essential to
searching for the appropriate information
(c) Felix Redmill, 2011
CERN, May '11
70
BOTTOM-UP ANALYSIS
(Herald of Free Enterprise)
Bowsun asleep in his cabin when ship is due to depart

Bow doors not closed

Ship puts to sea with bow doors open

Water enters car deck

As ship rolls, water rushes to one side

Ship capsizes

Lives lost
(c) Felix Redmill, 2011
CERN, May '11
71
TOP-DOWN ANALYSIS
(Herald of Free Enterprise)
Ship puts to sea with bow doors open
Bosun did not close doors
Bosun not available
to close doors
Bosun not
on ship
Bosun on board
but not at station
Bosun asleep
in cabin
(c) Felix Redmill, 2011
Problem with doors
& bosun can’t close them
Problem
with closing
mechanism
Door or
hinge
problem
Bosun in
bar
CERN, May '11
Problem
with power
supply
72
RISK MANAGEMENT PROCESS
• Define scope of study
• Identify the hazards
• Analyse the hazards to determine the risks they pose
• Assess risks against tolerability criteria
• Take risk-management decisions and actions
• It may also be appropriate to carry out emergency
planning and prepare for the unexpected
– If so, we need to carry out rehearsals
(c) Felix Redmill, 2011
CERN, May '11
73
FOUR STAGES OF RISK ANALYSIS
(But be careful with vocabulary)
• Definition of scope
– Define the objectives and scope of the study
• Hazard identification
– Define hazards and hazardous events
• Hazard analysis
– Determine the sequences leading to hazardous events
– Determine likelihood and consequences of hazardous
events
• Risk assessment
– Assess tolerability of risks associated with hazardous
events
(c) Felix Redmill, 2011
CERN, May '11
74
DEFINITION OF SCOPE
• Types of risks to be studied
– e.g. safety, security, financial
• Risks to whom or what
– e.g. employees, all people, environment, the company,
the mission
• Study boundary
– Plant boundary
– Region
• Admissible sources of information
– e.g. stakeholders (which?), experts, public
(c) Felix Redmill, 2011
CERN, May '11
75
SCOPE OF STUDY
• How will the results be used?
– What questions are to be answered?
– What decisions are to be made?
• What accuracy and confidence are required?
• Define study parameters
– What effort to be invested?
– What budget afforded?
– How much time allowed?
(c) Felix Redmill, 2011
CERN, May '11
76
OUTCOME SUBJECTIVELY INFLUENCED
• Defining scope is subjective
– Involves judgement
– Includes bias
– Can involve manipulation
• Scope definition
– Influences the nature and direction of the analysis
– Is a predisposing factor on its results
(c) Felix Redmill, 2011
CERN, May '11
77
VOCABULARY - CHOICE OF TERMS
• Same term means different things to different people
• Same process given different titles
• There is no internationally agreed vocabulary
• Even individuals use different terms for the same process
• Beware: ask what others mean by their terms
• Have a convention and define it
(c) Felix Redmill, 2011
CERN, May '11
78
SOME TERMS IN USE
Hazard identification,
Risk identification
Risk analysis,
Risk
assessment,
Risk
m anagement
Hazard analysis,
Risk analysis
Risk assessm ent,
Risk evaluation
Risk mitigation,
Risk reduction,
Risk managem ent
(c) Felix Redmill, 2011
CERN, May '11
79
AN APPROPRIATE CONVENTION?
Scope definition
Risk
analysis
Hazard identification
Hazard analysis
Risk assessment
Risk com munication
Risk m itigation
Risk
management
Em ergency planning
(c) Felix Redmill, 2011
CERN, May '11
80
HAZARD AND RISK ANALYSIS
• Obtaining appropriate information, from the most
appropriate sources, and analyzing it, while making
assumptions that are recognized, valid, and as few as
possible
(c) Felix Redmill, 2011
CERN, May '11
81
HAZARD IDENTIFICATION
• The foundation of risk analysis
- Identify the hazards (what could go wrong)
- Deduce their causes
- Determine whether they could lead to undesirable
outcomes
• Knowing chains of cause and effect facilitates decisions
on where to take corrective action
• But many accidents are caused by unexpected
interactions rather than by failures
(c) Felix Redmill, 2011
CERN, May '11
82
HAZARD ID METHODS
• Checklists
• Brainstorming
• Expert judgement
• What-if analysis
• Audits and reports
• Site inspections
• Formal and informal staff interviews
• Interviews with others, such as customers, visitors
• Specialised techniques
(c) Felix Redmill, 2011
CERN, May '11
83
WHAT WE FIND DEPENDS ON WHERE WE
LOOK
• We don’t find hazards where we don’t look
• We don’t look because
– We don’t think of looking there
– We don’t know of that place
– We assume there are no hazards there
– We assume that the risks are small
• Must take a methodical approach
– And be thorough
(c) Felix Redmill, 2011
CERN, May '11
84
CRACK IN “TIRELESS” COOLING SYSTEM
(c) Felix Redmill, 2011
CERN, May '11
85
EXTRACT FROM ‘FILE ON 4’ INTERVIEW (12/12/00)
•
•
•
•
•
•
•
•
•
•
(O'Halloran) For the Navy Captain Hurford concedes that the possibility that this critical
section of pipework might fail was never even considered in the many years that these 12
submarines of the Swiftsure and Trafalgar classes have been in service.
(Hurford) "This component was analysed against its duty that it saw in service and was
supposed never to crack and so the fact that this crack had occurred in this component in the
way that it did and caused a leak before we had detected it, is a serious issue.”
(O'Halloran) How big a question mark does this place over your general risk probability
assumptions … about the whole working of one of these nuclear reactors.
(Hurford) "It places a question on the surveillance that we do when the submarines are in
refit and maintenance, unquestionably”
(O'Halloran) How long have these various submarines been in service ?
(Hurford) "The oldest of the Swiftsure class came into service in the early seventies”
(O'Halloran) So has this area of the pipework ever been looked at in any of the submarines,
the 12 hunter killer submarines now in service ?
(Hurford) "No it hasn't, because the design of the component was understood and the
calculations showed and experience showed that there would be no problem.”
(O'Halloran) But the calculations were wrong ?
(Hurford) "Clearly there is something wrong with that component that caused the crack and
we don't know if it was the calculations or whether it was the way it was made and that what
is being found out in the analysis at the moment"
(c) Felix Redmill, 2011
CERN, May '11
86
REPEAT FORMAL HAZARD ID
• Hazard identification should be repeated
when new information becomes available and when
significant change is proposed. e.g.
– At the concept stage
– When a design is available
– When the system has been built
– Prior to system or environmental change during
operation
– Prior to decommissioning
(c) Felix Redmill, 2011
CERN, May '11
87
HAZARD ANALYSIS
• Analyse the identified hazards to determine
– Potential consequences
 Worst credible
 Most likely
– Ways in which they could lead to undesirable outcomes
– Likelihood of undesirable outcomes
• Sometimes rough estimates are adequate
• Sometimes precision is essential
– Infusion pump (too much: poison; too little: no cure)
– X-ray dose (North Stafford)
(c) Felix Redmill, 2011
CERN, May '11
88
NOTES ON HAZARD ANALYSIS
• Two aspects of a hazard: likelihood and consequence
• Consequence is usually closely related to system goals
– So risk reduction may focus on reduction of likelihood
• Usually numerous hazards associated with a system
• Analysis is based on collecting and deriving appropriate
information
• Pedigree of information depends on pedigree of its source
• The question of confidence should always be raised
• Likelihood and consequences may be expressed
quantitatively or qualitatively
(c) Felix Redmill, 2011
CERN, May '11
89
NOTES ON HAZARD ANALYSIS - 2
• For random events (such as hardware component failure)
– Historic failure frequencies may be known
– Probabilistic hazard analysis may be possible
• For systematic events (such as software failures)
– History is not an accurate predictor of the future
– Qualitative hazard analysis is necessary
• When low event frequencies (e.g. 10-6 per year) are desired,
confidence in figures must be low
(c) Felix Redmill, 2011
CERN, May '11
90
THE NEED FOR QUALITATIVE ANALYSIS
AND ASSESSMENT
• When failures and hazardous events are random
– Historic data may exist
– Numerical analysis may be possible
• When failures and hazardous events are systematic, or
historic data do not exist
– Qualitative analysis and assessment are necessary
• Example of qualitative techniques is the risk matrix
(c) Felix Redmill, 2011
CERN, May '11
91
HAZARD ANALYSIS TECHNIQUES
• FMEA analyses the effects of failures
• FMECA analyses the risks attached to failures
• HAZOP analyses both causes and effects
• Event tree analysis (ETA) works forward from identified
hazards or events (e.g. component failures) to determine
their consequences
• Fault tree analysis (FTA) works backwards from identified
hazardous events to determine their causes
(c) Felix Redmill, 2011
CERN, May '11
92
USING THE RESULTS
• The purpose is to derive reliable information on which to
base decisions on risk-management action
• ‘Derive’ - by carrying out risk analysis
• ‘Reliable’ - be thorough as appropriate
• ‘Information’ - the key to decision-making
• ‘Decisions’ - risk analysis is not carried out for its own sake
but to inform decisions (usually made by others)
• Hazard identification and analysis may be carried out
concurrently
(c) Felix Redmill, 2011
CERN, May '11
93
RISK ASSESSMENT
• To determine the tolerability of analysed risks
– So that risk-management decisions can be taken
• Need tolerability criteria
• Tolerability differs according to circumstance
– e.g. medical
• Tolerability differs in time
– e.g. nuclear
• Tolerability differs according to place
– e.g. oil companies treatment of staff and environment in
different countries
(c) Felix Redmill, 2011
CERN, May '11
94
TOLERABLE RISK
• Risk accepted in a given context based on the current
values of society
• Not trivial to determine
• Differs across industry sectors
• May change with time
• Depends on perception
• Should be determined by discussion among parties,
including
– Those posing the risks
– Those to be exposed to the risks
– Other stakeholders, e.g. regulators
(c) Felix Redmill, 2011
CERN, May '11
95
THE HSE’S ALARP PRINCIPLE
(c) Felix Redmill, 2011
CERN, May '11
96
RISK TOLERABILITY JUDGEMENT
• In the case of machinery
– We may know what could go wrong (uncertainty is low)
– It may be reversible (can fix it after one accident)
• In the case of BSE or genetically modified organisms
– The risk may only be suspected (uncertainty is high)
– It may be irreversible
• Moral values influence the judgement of how much is enough
• If we don’t take risks we don’t make progress
(c) Felix Redmill, 2011
CERN, May '11
97
HAZARD AND RISK ANALYSIS
TECHNIQUES
(c) Felix Redmill, 2011
CERN, May '11
98
TECHNIQUES
Techniques support risk analysis
– They should not govern it
(c) Felix Redmill, 2011
CERN, May '11
99
TECHNIQUES TO BE CONSIDERED
•
•
•
•
•
•
•
•
Failure (fault) modes and effects analysis (FMEA)
Failure (fault) modes, effects and criticality analysis (FMECA)
Hazard and operability studies (HAZOP)
Event tree analysis (ETA)
Fault tree analysis (FTA)
Risk matrices
Human reliability analysis (HRA - mentioned only)
Preliminary hazard analysis (PHA)
(c) Felix Redmill, 2011
CERN, May '11
100
A SIMPLE CHEMICAL PLANT
P1
Fluid A
V1
V2
Vat R
P2
Fluid B
V3
(c) Felix Redmill, 2011
CERN, May '11
101
FAILURE MODES AND EFFECTS ANALYSIS
• Usually qualitative investigation of the modes of failure of
individual system components
• Components may be at any level (e.g. basic, sub-system)
• Components are treated as black boxes
• For each failure mode, FMEA investigates
– Possible causes
– Local effects
– System-level effects
• Corrective action may be proposed
• Best done by a team of experts with different viewpoints
(c) Felix Redmill, 2011
CERN, May '11
102
FAILURE MODES AND EFFECTS ANALYSIS
• Boundary of study must be clearly defined
• Does not usually find the effects of
– Multiple faults or failures
– Failures caused by communication and interactions
between components
– Integration of components
– Installation
(c) Felix Redmill, 2011
CERN, May '11
103
EXAMPLE FMEA OF A SIMPLE CHEMICAL
PROCESS
Study
No.
1
It em
Pump
P1
2
3
Failure
mode
Fails to
st art
Burns out
Valve
V1
4
Sticks
closed
Sticks
open
Possible
causes
1. No power
2. Burnt out
1. Loss of
lubricant
2. Excessive
t emperatur e
1. No power
2. Jammed
1. No power
2. Jammed
(c) Felix Redmill, 2011
CERN, May '11
Local
effe cts
Fluid
does
flow
Fluid
does
flow
A
not
A
not
Fluid A
cannot
flow
Cannot
st op
flow of
Fluid A
Syst emlevel
effe cts
Excess of
Fluid B in
Vat R
Excess of
Fluid B in
Vat R
Monitor
pump
operat ion
Add alarm
to pump
monito r
Excess
Fluid B
Vat R
Danger
excess
Fluid A
Vat R
Monitor
Valve
operat ion
Int roduce
addit ional
valve in
series
of
in
of
of
in
Proposed
correct ion
104
FAILURE MODES, EFFECTS AND
CRITICALITY ANALYSIS
• FMECA = FMEA + analysis of the criticality of each failure
• Criticality in terms of risk or one of its components
– i.e. severity and probability or frequency of occurrence
– Usually quantitative
• Purpose: to identify the failure modes of highest risk
– This facilitates prioritization of corrective actions,
focusing attention and resources first where they will
have the greatest impact on safety
• Recording: add one or two columns to a FMEA table
(c) Felix Redmill, 2011
CERN, May '11
105
HAZARD AND OPERABILITY STUDIES
(HAZOP)
• The methodical study of a documented representation of a
system, by a managed team, to identify hazards and
operational problems.
• Correct system operation depends on the attributes of the
items on the representation remaining within design intent.
Therefore, studying what would happen if the attributes
deviated from design intent should lead to the identification
of the hazards associated with the system's operation. This
is the principle of HAZOP.
• 'Guide words' are used to focus the investigation on the
various types of deviation.
(c) Felix Redmill, 2011
CERN, May '11
106
A GENERIC SET OF GUIDE WORDS
Guide word
Meaning
No
The complete negat ion of t he design int ent ion. No part of t he
int ent ion is achieved, but nothing else happens.
More
A quant it at ive increase.
Less
A quant it at ive decrease.
As well as
A qualit at ive increase. All t he design int ent ion is achieved, t ogether
wit h somet hing addit ional.
Part of
A qualit at ive decrease. Only part of t he design int ent ion is achieved.
Reverse
The logical opposit e of t he design int ent ion.
Other t han
A complete subst it ut ion where no part of t he design int ent ion is
achieved but somet hing diff erent occurs.
Early
The design int ent ion occurs earlier in t ime t han int ended.
Late
The design int ent ion occurs lat er in t ime t han int ended.
Before
The design int ent ion occurs earlier in a sequence t han int ended.
Af t er
The design int ent ion occurs earlier in a sequence t han int ended.
(c) Felix Redmill, 2011
CERN, May '11
107
HAZOP OVERVIEW
Introductions
Presentation of design
representation
Examine
design representation
methodically
Possible
deviation from design intent
?
No
Yes
Examine
consequences
and causes
Document
results
Define follow-up
work
No
T ime up, or
completed study?
Yes
Agree documentation
Sign off meeting
(c) Felix Redmill, 2011
CERN, May '11
108
HAZOP SUMMARY
• HAZOP is a powerful technique for hazard identification and analysis
• It requires a team of carefully chosen members
• It depends on planning, preparation, and leadership
• A study usually requires several meetings
• Study proceeds methodically
• Guide words are used to focus attention
• Outputs are the identified hazards, recommendations,
questions
(c) Felix Redmill, 2011
CERN, May '11
109
EVENT TREE ANALYSIS
• Starts with an event that might affect safety
• Follows a cause-and-effect chain to the system-level
consequence
• Does not assume that an event is hazardous
• Includes both operational and fault conditions
• Each event results in a branch, so N events = 2N branches
• If event probabilities can be derived
– They may be assigned to the branches of the tree
– The probability of the final events may be calculated
(c) Felix Redmill, 2011
CERN, May '11
110
A SIMPLE EVENT TREE
Valve
operates
Valve
monitor
O.K.
Alarm
relay O.K.
Claxon
O.K.
Operator
responds
Yes
Yes
Outcome
Safe
outcome
No
Yes
No
Yes
No
Unsafe
outcomes
No
No
(c) Felix Redmill, 2011
CERN, May '11
111
REFINED EVENT TREE
Valve
operates
Valve
monitor
functions
Alarm
relay
operates
Claxon
sounds
Yes
Yes
Operator
responds
Yes
Yes
No
No
(c) Felix Redmill, 2011
CERN, May '11
112
EVENT TREE: ANOTHER EXAMPLE
Fire
starts
Fire
spreads
quickly
Sprinkler
fails to
work
People
cannot
escape
Yes (P=0.4)
Yes (P=0.2)
No (P=0.6)
Yes (P=0.1)
No (P=0.8)
Resulting
event
Multiple
fatalities
Damage
and loss
Fire
controlled
Yes
No (P=0.9)
(c) Felix Redmill, 2011
Fire
contained
CERN, May '11
113
BOTTOM-UP ANALYSIS
(Herald of Free Enterprise)
Bowsun asleep in his cabin when ship is due to depart

Bow doors not closed

Ship puts to sea with bow doors open

Water enters car deck

As ship rolls, water rushes to one side

Ship capsizes

Lives lost
(c) Felix Redmill, 2011
CERN, May '11
114
TOP-DOWN ANALYSIS
(Herald of Free Enterprise)
Ship puts to sea with bow doors open
Bosun did not close doors
Bosun not available
to close doors
Bosun not
on ship
Bosun on board
but not at station
Bosun asleep
in cabin
(c) Felix Redmill, 2011
Problem with doors
& bosun can’t close them
Problem
with closing
mechanism
Door or
hinge
problem
Bosun in
bar
CERN, May '11
Problem
with power
supply
115
FAULT TREE ANALYSIS
• Starts with a single 'top event' (e.g., a system failure)
• Combines the chains of cause and effect which could lead
to the top event
• Next-level causes are identified and added to the tree and
this is repeated until the most basic causes are identified
• Causal relationships are defined by AND and OR gates
• One lower-level event may cause several higher-level
events
• Examples of causal events: component failure, human
error, software bug
• Each top-level undesired event requires its own fault tree
• Probabilities may be attributed to events, and from these
the probability of the top event may be derived
(c) Felix Redmill, 2011
CERN, May '11
116
EXAMPLE FTA OF SIMPLE CHEMICAL
PROCESS
(c) Felix Redmill, 2011
CERN, May '11
117
THE PRINCIPLE OF THE PROTECTION SYSTEM
Required event
frequency
10 -7 per hour
AND
Dangerous failure
frequency of
equipt.: 10 - 3 / hour
(c) Felix Redmill, 2011
Reliability of safety
function
10 -4 / hour
CERN, May '11
118
COMPLEMENTARITY OF TECHNIQUES
• Compare results of FMEA with low-level causes from FTA
• Carry out HAZOP on a sub-system identified as risky by a
high-level FMEA
• Carry out ETA on low-level items identified as risky by FTA
(c) Felix Redmill, 2011
CERN, May '11
119
A RISK MATRIX
(An example of qualitative risk analysis)
Likelihood
or
Frequency
Consequence
Negligible
Moderate
High
Catastrophic
High
Medium
Low
(c) Felix Redmill, 2011
CERN, May '11
120
RISKS POSED BY IDENTIFIED HAZARDS
Likelihood
or
Frequency
Consequence
Negligible
High
Medium
Moderate
Catastrophic
H2
H1, H5
H6
Low
(c) Felix Redmill, 2011
High
H4
H3
CERN, May '11
121
A RISK CLASS MATRIX
(example only)
Likelihood
or
Frequency
Consequence
Negligible
Moderate
High
Catastrophic
High
C
B
A
A
Medium
D
C
B
A
Low
D
D
C
B
(c) Felix Redmill, 2011
CERN, May '11
122
TOLERABILITY OF ANALYSED RISKS
• Refer the cell of each analysed hazard in the risk matrix to the
equivalent cell in the risk class matrix to determine the class
of the analysed risks
– So, Hazard 1 poses a D Class risk, Hazard 2 a B, etc.
• Risk class
– Defines the (relative) importance of risk reduction
– Provides a means of prioritising the handling of risks
– Can be equated to a defined type of action
• Risk class gives no indication of time criticality
– This must be derived from an understanding of the risk
• What is done to manage risks depends on circumstances
(c) Felix Redmill, 2011
CERN, May '11
123
THOUOGHTS ON THE USE OF TECHNIQUES
• Techniques are intended to support what you have decided
to do
– They should not define what you do
• Each is useful on its own for a given purpose
– Thorough analysis requires a combination of techniques
• Risk analysis is not achieved in a single activity
– It should be continuous throughout a system’s life
• Early analysis (PHA) is particularly effective
• Techniques may be used to verify each other’s results
• And to expand on them and provide further detail
(c) Felix Redmill, 2011
CERN, May '11
124
IMPORTANCE OF EARLY ANALYSIS
Root causes
1000s
Hazards
<100
Accidents
<20
• Too many causes to start with bottom-up analysis
– Effort and financial cost would be too great
– Would lead to over-engineering for small risks
– Would miss too many important hazards
• Fault trees commence with accidents or hazards
• Carry out Preliminary Hazard Analysis early in a project
(c) Felix Redmill, 2011
CERN, May '11
125
PRELIMINARY HAZARD ANALYSIS
• At the ‘Objectives’ or early specification stage
– Potential then for greatest safety effectiveness
• Take a ‘system’ perspective
• Consider historic understanding of such systems
• Address differences between this system and historic
systems (novel features)
• Consider such matters as: boundaries, operational intent,
physical operational environment, assumptions
• Identify accident types, potential accident scope and
consequences
• Identify system-level hazards
• If a checklist is used, review it for obsolescence
• Create a new checklist for future use
(c) Felix Redmill, 2011
CERN, May '11
126
HUMAN RELIABILITY ASSESSMENT
• Human components of systems are unreliable
• Hazards posed by them should be included in risk analysis
• Several Human Reliability Assessment (HRA) techniques
have been developed for the purpose
– These will be covered later in the degree course
(c) Felix Redmill, 2011
CERN, May '11
127
RISKS POSED BY HUMANS
• We need to
– Extend risk analysis to include HRA
– Pay more attention to ergonomic considerations in
design
– Consider human cognition in interface design
– Produce guidelines on human factors in safety
• Safety is multi-dimensional, so take an interdisciplinary
approach
– Include ergonomists and psychologists in development
and operational teams
(c) Felix Redmill, 2011
CERN, May '11
128
RISKS POSED BY SENIOR MANAGERS
• Senior managers (should)
– Define safety policy
– Make strategic safety decisions
– Provide leadership in safety (or not)
– Define culture by design or by their behaviour
• They predispose an organisation to safety or accident
• Accident inquiry reports show that the contribution of
senior management to accident causation is considerable
• Yet their contributions to risk are not included in risk
analyses
• Risk analyses therefore tend to be optimistic
• Risk analyses need to include management failure
• Management should pay attention to management failure
(c) Felix Redmill, 2011
CERN, May '11
129
RISK ANALYSIS SUMMARY
• Risk analysis is the cornerstone of modern safety engineering
and management
• It can be considered as a four-staged process
• The first stage, hazard identification, is the basis of all further
work - risks not identified are not analysed or mitigated
• All four stages include considerable subjectivity
• Subjectivity, in the form of human creativity and judgement, is
essential to the effectiveness of the process
• Subjectivity also introduces bias and error
• HRA techniques need to be included in risk analyses
• We need techniques to include management risks
• Often the greatest value of risk analysis is having to do it
(c) Felix Redmill, 2011
CERN, May '11
130
SAFETY INTEGRITY LEVELS
(SILs)
(c) Felix Redmill, 2011
CERN, May '11
131
SAFETY AS A WAY OF THINKING
• Reliability is not a guarantee of safety
• If risk is too great, it must be reduced
(the beginning of 'safety thinking')
• But how do we know if the risk is too great?
– Carry out risk analysis
(safety engineering enters software and systems
engineering)
• But even if the risk is high, does it need to be reduced?
– Determine what level of risk is tolerable
(this introduces the concept of safety management)
(c) Felix Redmill, 2011
CERN, May '11
132
SAFETY INTEGRITY
• If risk is not tolerable it must be reduced
• High-level requirement of safety function is 'to reduce the
risk'
• Analysis leads to the functional requirements
• The safety function becomes part of the overall system
– Safety depends on it
• So, will it reduce the risk to (at least) a tolerable level?
• We try to ensure that it does by defining the reliability with
which it performs its safety function
– In IEC 61508 in terms of Pr. of dangerous failure
• This is referred to as 'safety integrity'
(c) Felix Redmill, 2011
CERN, May '11
133
SAFETY INTEGRITY IS A TARGET
REQUIREMENT
• Safety integrity sets a target for the maximum tolerable rate of
dangerous failures of a safety function
– (This is a ‘reliability-type’ attribute)
• Example: S should not fail dangerously more than once in 7
years
• Example: There should be no more than 1 dangerous failure in
1000 demands
• If the rate of dangerous failures of the safety function can be
measured accurately, achievement of the defined safety
integrity may be claimed
(c) Felix Redmill, 2011
CERN, May '11
134
WHY SAFETY INTEGRITY LEVELS?
• A safety integrity target could have an infinite number of
values
• It is practical to divide the total possible range into bands
(categories)
• The bands are 'safety integrity levels' (SILs)
– In IEC 61508 there are four
• N.B. Interesting that the starting point was that
safety and reliability are not synonymous
and the end point is the reliance of safety
on a reliability-type measure (rate of dangerous failure)
(c) Felix Redmill, 2011
CERN, May '11
135
IEC 61508 SIL VALUES (IEC 61508)
Safety
Integrity
Level
Low Demand Mode of
Operation
(Pr. of failure to perform its
safety functions on demand)
Continuous/High-demand Mode
of Operation
(Pr. of dangerous failure per
hour)
4
>= 10-5 to 10-4
>= 10-9 to 10-8
3
>= 10-4 to 10-3
>= 10-8 to 10-7
2
>= 10-3 to 10-2
>= 10-7 to 10-6
1
>= 10-2 to 10-1
>= 10-6 to 10-5
(c) Felix Redmill, 2011
CERN, May '11
136
RELATIONSHIP (IN IEC 61508) BETWEEN
LOW-DEMAND AND CONTINUOUS MODES
• Low-demand mode:
'Frequency of demands ... no greater than one per year'
• 1 year = 8760 hours (approx. 104)
• Continuous and low-demand figures related by factor of 104
• Claims for a low-demand system must be justified
(c) Felix Redmill, 2011
CERN, May '11
137
APPLYING SAFETY INTEGRITY TO
DEVELOPMENT PROCESS
• When product integrity cannot be measured with confidence
(particularly when systematic failures dominate)
– The target is related, in IEC 61508, to the development
process
• Processes are equated to safety-integrity levels
• But
– Such equations are judgemental
– Process may be an indicator of product, but not an
infallible one
(c) Felix Redmill, 2011
CERN, May '11
138
IEC 61508 DEFINITIONS
• Safety integrity:
Probability of a safety-related system satisfactorily
performing the required safety functions under all the
stated conditions within a stated period of time
• Safety integrity level:
Discrete level (one out of a possible four) for specifying the
safety integrity requirements of the safety functions to be
allocated to the E/E/PE safety-related systems, where SIL 4
has the highest level of safety integrity and SIL 1 the lowest
(c) Felix Redmill, 2011
CERN, May '11
139
SIL IN TERMS OF RISK REDUCTION (IEC 61508)
• The necessary risk reduction effected by a safety function
• Its functional requirements define what it should do
• Its safety integrity requirements define its tolerable rate of
dangerous failure
(c) Felix Redmill, 2011
CERN, May '11
140
THE IEC 61508 MODEL OF RISK
REDUCTION
Utility + risks
EUC
Control
system
Safety functions
(c) Felix Redmill, 2011
Protection
system
Safety functions
CERN, May '11
141
PRINCIPLE OF THE PROTECTION SYSTEM
Event frequency
10 -7
AND
EUC dangerous
failure frequency
Reliability of
safety function
10 -3
10 -4
(c) Felix Redmill, 2011
10 -4
10 -3
CERN, May '11
142
BEWARE
• Note that a ‘protection system’ claim requires total
independence of the safety function from the protected
function
(c) Felix Redmill, 2011
CERN, May '11
143
EACH HAZARD CARRIES A RISK
Residual
risk 1
T olerable
level of
risk 1
Risk
2
T olerable
level of
risk 2
Necessary reduction of risk 1
Risk
1
Increasing
risk
Actual reduction of risk 1
(c) Felix Redmill, 2011
CERN, May '11
144
EXAMPLE OF A SAFETY FUNCTION
•
•
The target SIL applies to the entire safety instrumentation system
– Sensor, connections, hardware platform, software, actuator, valve,
data
All processes must accord with the SIL
– Validation, configuration management, proof checks, maintenance,
change control and revalidation ...
(c) Felix Redmill, 2011
CERN, May '11
145
PROCESS OF SIL DERIVATION
• Hazard identification
• Hazard analysis
• Risk assessment
– Resulting in requirements for risk reduction
• Safety requirements specification
– Functional requirements
– Safety integrity requirements
• Allocation of safety requirements to safety functions
– Safety functions to be performed
– Safety integrity requirements
• Allocation of safety functions to safety-related systems
– Safety functions to be performed
– Safety integrity requirements
(c) Felix Redmill, 2011
CERN, May '11
146
THE IEC 61508 SIL PRINCIPLE
(c) Felix Redmill, 2011
CERN, May '11
147
BUT NOTE
• IEC 61508 emphasises process evidence but does not
exclude the need for product or analysis evidence
• It is impossible for process evidence to be conclusive
• It is unlikely that conclusive evidence can be derived for
complex systems in which systematic failures dominate
• Evidence of all types should be sought and assembled
(c) Felix Redmill, 2011
CERN, May '11
148
SIL ALLOCATION - 1 (from IEC 61508)
• Functions within a system may be allocated different SILs
• The required SIL of hardware is that of the software safety
function with the highest SIL
• For a safety-related system that implements safety
functions of different SILs, the hardware and all the
software shall be of the highest SIL unless it can be shown
that the implementations of the safety functions of the
different SILs are sufficiently independent
• Where software is to implement safety functions of different
SILs, all the software shall be of the highest SIL unless the
different safety functions can be shown to be independent
(c) Felix Redmill, 2011
CERN, May '11
149
SIL ALLOCATION - 2
(from IEC 61508)
• Where a safety-related system is to implement both safety
and non-safety functions, all the hardware and software
shall be treated as safety-related unless it can be shown
that the implementation of the safety and non-safety
functions is sufficiently independent (i.e. that the failure of
any non-safety-related functions does not affect the safetyrelated functions). Wherever practicable, the safety-related
functions should be separated from the non-safety-related
functions
(c) Felix Redmill, 2011
CERN, May '11
150
SIL — ACHIEVED OR TARGET
RELIABILITY?
• SIL is initially a statement of target rate of dangerous failures
• There may be reasonable confidence that achieved reliability =
or > target reliability (that the SIL has been met), when
– Simple design
– Simple hardware components with known fault histories in
same or similar applications
– Random failures
• But there is no confidence that the target SIL is met when
– There is no fault history
– Systematic failures dominate
• Then, IEC 61508 focuses on the development process
(c) Felix Redmill, 2011
CERN, May '11
151
APPLICATION OF SIL PRINCIPLE
• IEC 61508 is a meta standard, to be used as the basis for
sector-specific standards
• The SIL principle may be adapted to suit the particular
conditions of the sector
• An example is the application of the principle in the Motor
Industry guideline
(c) Felix Redmill, 2011
CERN, May '11
152
EXAMPLE OF SIL BASED ON
CONSEQUENCE
(Motor Industry)
• In the Motor Industry Software Reliability Association (MISRA)
guidelines (1994), the allocation of a SIL is determined by 'the
ability of the vehicle occupants … to control the situation
following a failure'
• Steps in determining an integrity level:
a) List all hazards that result from all failures of the system
b) Assess each failure mode to determine its controllability
category
c) The failure mode with the highest associated controllability
category determines the integrity level of the system
(c) Felix Redmill, 2011
CERN, May '11
153
MOTOR INDUSTRY EXAMPLE
Controllability
category
Acceptable failure rate
Integrity
level
Uncontrollable
Extremely improbable
4
Difficult to control
Very remote
3
Debilitating
Remote
2
Distracting
Unlikely
1
Nuisance only
Reasonably possible
0
(c) Felix Redmill, 2011
CERN, May '11
154
SIL CAN BE MISLEADING
• Different standards derive SILs differently
• SIL is not a synonym for overall reliability
– Could lead to over-engineering and greater costs
• When SIL claim is based only on development process
– What were competence, QA, safety assessment, etc.?
• Glib use of the term ‘SIL’
– No certainty of what is meant
– Is the claim understood by those making it?
– Be suspicious until examining the evidence
• SIL X may be intended to imply use in all circumstances
– But safety is context-specific
• SIL says nothing about required attributes of the system
– Must go back to the specification to identify them
(c) Felix Redmill, 2011
CERN, May '11
155
THREE QUESTIONS
• Does good process necessarily lead to good product?
• Instead of using a safety function, why not simply improve
the basic system (EUC)?
• Can the SIL concept be applied to the basic system?
(c) Felix Redmill, 2011
CERN, May '11
156
DOES GOOD PROCESS LEAD TO GOOD PRODUCT?
• Adhering to “good” process does not guarantee reliability
– Link between process and product is not definitive
• Using a process is only a start
– Quality, and safety, come from professionalism
• Not only development processes, but also operation,
maintenance, management, supervision, etc., throughout the
life cycle
• Not only must development be rigorous but also the safety
requirements must be correct
– And requirements engineering is notoriously difficult
• Meeting process requirements offers confidence, not proof
(c) Felix Redmill, 2011
CERN, May '11
157
WHY NOT SIMPLY IMPROVE THE BASIC
SYSTEM?
•
•
•
•
IEC 61508 bases safety on the addition of safety functions
This assumes that EUC + control system are fixed
Why not reduce risk by making EUC & control system safer?
This is an iterative process
– Then, if all risks are tolerable no safety functions required
– But (according to IEC 61508) claim cannot be lower than
10-5 dangerous failures/hour (SIL 1)
• If there comes a time when we can't (technological) or won't
(cost) improve further
– We must add safety functions if all risks are not tolerable
• We are then back to the standard as it is
(c) Felix Redmill, 2011
CERN, May '11
158
CAN THE SIL CONCEPT BE APPLIED TO THE
BASIC SYSTEM?
• Determine its tolerable rate of dangerous failure and translate
this into a SIL (as in the MISRA guidelines)
• But note the limit on the claim that can be made (10-5
dangerous failures per hour) to satisfy IEC 61508
(c) Felix Redmill, 2011
CERN, May '11
159
DOES MEETING THE SIL GUARANTEE SAFETY?
• Derivation of a SIL is based on hazard and risk analyses
• Hazard and risk analyses are subjective processes
• If the derived risk value is an accurate representation of the
risk AND the selected level of risk tolerability is not too high
THEN meeting the SIL may offer a good chance of safety
• But these criteria may not hold
• Also, our belief that the SIL has been met may not be justified
(c) Felix Redmill, 2011
CERN, May '11
160
SIL SUMMARY
• The SIL concept specifies the tolerable rate of dangerous
failures of a safety-related system
– It defines a “safety-reliability” target
• Evidence of achievement comes from product and process
– IEC 61508 SIL defines constraints on development
processes
• Different standards use the concept differently
– SIL derivation may be sector-specific
• It can be misleading in numerous ways
• The SIL concept is a tool
– It should be used only with understanding and care
(c) Felix Redmill, 2011
CERN, May '11
161
Download