5.0 Instrument calibrations

advertisement
DRAFT ANSI N323C-200x
1
2
3
4
AMERICAN NATIONAL STANDARD
5
6
7
8
9
10
American National Standard
Radiation Protection Instrumentation
Test and Calibration –
Air Monitoring Instruments
11
12
13
FINAL DRAFT – July 2008
N42 Comment Resolution Copy
14
15
16
17
Sponsor
National Committee on Radiation Instrumentation, N42
18
19
20
21
Accredited by the
American National Standards Institute
22
23
24
25
Secretariat
Institute of Electrical and Electronics Engineers, Inc.
26
27
28
Approved (date to be determined)
American National Standards Institute
29
30
31
32
Abstract- This standard establishes test and calibration requirements for air
monitoring instruments used for the detection and measurement of airborne
radioactive substances.
1
1
2
3
Key words: radioactive air monitoring, calibration and tests.
2
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
Participants
At the time it approved this standard, the Accredited Standards Committee on Radiation
Instrumentation, N42, had the following membership:
Michael P. Unterweger, Chair
Louis Costrell, Deputy Chair
William Ash, Administrative Secretary
Organization Represented .......................................................................... Name of Representative
Canberra ..................................................................................................................Markku Koskelo
Chew, M.H. ................................................................................................................. Jack M. Selby
Commerce Dept, U.S. NIST ............................................................................. Michael Unterweger
........................................................................................................................... Louis Costrell (Alt.)
Department of Homeland Security............................................................................... Peter Shebell
Entergy-ANO .............................................................................................................. Ron Schwartz
Health Physics Society. .................................................................................................. Sandy Perle
Institute of Electrical & Electronics Engineers, Inc. ................................................... Louis Costrell
............................................................................................................................ Julian Forster (Alt.)
...................................................................................................................... Anthony Spurgin (Alt.)
................................................................................................................. Michael Unterweger (Alt.)
International Medcom .......................................................................................................Don Sythe
Lawrence Berkeley National Laboratory .............................................................. Edward J. Lampo
Lawrence Livermore National Laboratory ...................................................................Gary Johnson
NASA, GSFC ................................................................................................ R. Sachidananda Babu
U. S. Nuclear Regulatory Commission ....................................................................... Cynthia Jones
Nuclear Standards Unlimited .................................................................................. Al N. Tschaeche
Oak Ridge National Laboratory ............................................................................ Peter J. Chiaro, Jr
......................................................................................................................... Charles Britton (Alt.)
Ortec Corp. ........................................................................................................... Ronald M. Keyser
Pacific Northwest National Laboratory.................................................................... Richard Kouzes
Swinth Associates ............................................................................................... Kenneth L. Swinth
U.S. Army ............................................................................................................... Edward Groeber
U. S. Nuclear Regulatory Commission ....................................................................... Cynthia Jones
Members-At-Large........................................................................................................ Morgan Cox
.................................................................................................................................. Frank X. Masse
.......................................................................................................................... Joseph C. McDonald
.................................................................................................................................... Paul L. Phelps
.................................................................................................................................... Joseph Stencel
.................................................................................................................................... Lee J. Wagner
.................................................................................................................................... Ernesto. Corte
3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
At the time this standard was approved, Subcommittee N42.RPI had the following membership:
Morgan Cox, Co-Chair
Jack M. Selby, Co-Chair
Dru Carson
Peter J. Chiaro, Jr.
Jack Cooley
Leo Faust
Edward Groeber
Jerry Hiatt
Mark D. Hoover
Ron Keyser
J. C. McDonald
Robert Murphy
Cheryl Olson
Scott Rogers
Michael P. Unterweger
Ed Walker
Chuan-Fu Wu
At the time this standard was approved, the ANSI N323C Working Group had the following
members:
Michelle L. Johnson, Co-Chair
Mark D. Hoover, Co-Chair
L. Bryan Belvin
Brent Blunt
Morgan Cox
Stephen A. Epperson
Michael S. Ford
Robert J. Ford
John A. Glissmeyer
Robert A. Kellner
Timothy Martinson
4
George J. Newton
John C. Rodgers
Johnafred M. Thomas
James T. Voss
1
2
CONTENTS1.0
7
SCOPE
3
1.1
Applications of this standard
7
4
1.2
Special word usage
9
5
2.0
DEFINITIONS
9
6
3.0
PROGRAM ELEMENTS
12
7
3.1
Type testing
13
8
9
10
11
3.2
Acceptance testing
3.2.1
Physical inspection
3.2.2
General operations test
3.2.3
Radiological response tests
13
14
14
14
12
3.3
Initial calibration
14
13
3.4
Functional checks
14
14
15
16
3.5
Maintenance and recalibration
3.5.1
Maintenance
3.5.2
Recalibration
15
16
16
17
3.6
Performance tests
17
18
4.0
AIR FLOW RATE CALIBRATIONS
17
19
5.0
INSTRUMENT CALIBRATIONS
18
20
5.1
Pre-calibration inspection
18
21
5.2
As-found readings
19
22
5.3
Electronic calibration
19
23
24
25
26
27
5.4
Meters and chart recorders
5.4.1
Analog linear readout instruments
5.4.2
Analog logarithmic readout instruments
5.4.3
Digital readout instruments
5.4.4
Microprocessor-based instruments
19
19
20
20
21
28
5.5
Background response verification
21
29
5.6
Calibration of background subtraction
21
30
5.7
Alarm circuit verification
21
5
1
5.8
Radiological calibration
22
2
3
6.0 ADDITIONAL CALIBRATION REQUIREMENTS FOR PARTICULATE AND
NOBLE GAS MONITORS
22
4
6.1
Alpha and beta particulate monitors
22
5
6.2
Keep-alive sources
23
6
7
8
9
6.3
Noble gas monitors
6.3.1
Pre-calibration
6.3.2
Electronic calibration
6.3.3
Radiological calibration
23
23
23
23
10
7.0
CALIBRATION FOR SPECIAL CONDITIONS
24
11
8.0
FACILITIES, CONDITIONS, AND EQUIPMENT
25
12
8.1
Calibration facility
25
13
8.2
Test and calibration conditions
25
14
8.3
Calibration standards
26
15
8.4
Maintenance of standards
27
16
8.5
Check sources
28
17
9.0
DOCUMENTATION
28
18
9.1
Facility documentation
28
19
9.2
Instrument Documentation
29
20
10.0
21
22
APPENDIX A - SUMMARY OF A LIFE-CYCLE APPROACH TO INSTRUMENT
DEVELOPMENT AND APPLICATION
32
23
24
APPENDIX B - FLOW RATE METER CALIBRATIONS
34
REFERENCES
29
25
6
1
2
3
4
AMERICAN NATIONAL STANDARD
RADIATION PROTECTION INSTRUMENTATION
TEST AND CALIBRATION –
AIR MONITORING INSTRUMENTS
5
6
7
8
9
10
11
12
13
14
15
1.0
Scope
This standard establishes test and calibration requirements for air monitoring
instruments used for detection and measurement of airborne radioactive substances.
The appendices of this standard provides reference information.
The standard covers the following topics:

Test and calibration elements of a comprehensive air monitoring instrument
program;
16

Test and calibration requirements for all types of air monitoring instruments;
17

Additional requirements for specialized monitors;
18

Calibration for special conditions;
19

Required conditions, facilities, and equipment, and
20

Documentation.
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
1.1
Applications of this standard
This standard applies to the calibration of installed, portable, and personal air
monitoring instruments used for regulatory compliance, including continuous air
monitors for effluent, work place, and environmental applications. It may not be
universally applicable to all programs, especially those with more limited objectives
such as process control. When calibrating instrumentation used for objectives other
than regulatory compliance, the user of this standard should exercise judgment in the
application of these requirements and should document the monitoring objectives and
the reasons for any exceptions to the requirements of this standard. Relationship to
other standards
This standard shall be used in conjunction with ANSI N42.17B-1989 listed below. Many
items herein are identical to those in N42.17B, for example, the standard test conditions
of Table 2 in section 8.2.
The following national and international standards specify design and performance
issues, including when, where, and how air monitoring instruments shall, should, or may
be used to sample or monitor airborne radioactive substances. Such issues are
7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
addressed in this standard to the extent that they influence the ability to achieve proper
calibration.
ANSI N42.17B-1989, Performance Specifications for Health Physics Instrumentation –
Occupational Airborne Radioactivity Monitoring Instrumentation, specifies performance
criteria and testing procedures for instruments and instrument systems designed to
continuously sample and quantify concentrations of radioactivity in ambient air in the
workplace. It does not specify which instruments or systems are required, nor does it
address the specific locations or applications of such instruments.
ANSI N42.18-1974 (Reaffirmed 2004), Specification and Performance of On-Site
Instrumentation for Continuously Monitoring Radioactivity in Effluents, provides
recommendations for the selection of continuous monitors specific to the continuous
monitoring and quantification of radioactivity in effluents released to the environment
from nuclear facilities under normal operating conditions. It applies to continuous
monitors that measure normal releases, detect inadvertent releases, show general
trends, and annunciate radiation levels that have exceeded predetermined levels. It
does not cover emergency situations, sample extraction, laboratory analyses,
environmental monitoring, or process control.
ANSI N13.1-1999, Sampling and Monitoring Releases of Airborne Radioactive
Substances from the Stacks and Ducts of Nuclear Facilities, presents a comprehensive
approach to achieving an unbiased, representative sample from air streams by meeting
rigorous criteria for obtaining a well-mixed sample.
ANSI N320-1979 (Reaffirmed 2003), Performance Specifications for Reactor
Emergency Radiological Monitoring Instrumentation, addresses the essential
performance parameters of monitoring instruments used during an accident at nuclear
reactors. The instrument operating environment, operational characteristics, lower and
upper detection limits are also addressed. The general instrument locations inside the
reactor plant, at release points, and in the plant environs are addressed.
ANSI N42.22-1995 (Reaffirmed 2002), Traceability of Radioactive Sources to the
National Institute of Standards and Technology (NIST) and Associated Instrument
Quality Control, provides a description of the criteria necessary for manufacturers to
maintain and assure measurement traceability of radionuclides to NIST.
ANSI N42.23-1996 (Reaffirmed 2004), Measurement and Associated Instrumentation
Quality Assurance for Radioassay Laboratories, provides the framework to create a
national or an organizational NIST-traceable measurement quality assurance (MQA)
program.
ANSI N42.30-1999, Performance Specification for Tritium Monitors, provides
performance requirements for tritium monitors used for monitoring airborne tritium
radioactivity. It also provides set methods that can be used to establish performance
characteristics of a monitor.
8
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
IEC 60761-1 (2002) Equipment for Continuously Monitoring Radioactivity in Gaseous
Effluents-Part 1: General requirements. These five standards are self explanatory
based on their titles.
IEC 60761-2 (2002) Equipment for Continuously Monitoring Radioactivity in Gaseous
Effluents- Part 2: Specific requirements for radioactive aerosol monitors including
transuranic aerosols.
IEC 60761-3 (2002) Equipment for Continuously Monitoring Radioactivity in Gaseous
Effluents-Part 3: Specific requirements for radioactive noble gas monitors.
IEC 60761-4 (2002) Equipment for Continuously Monitoring Radioactivity in Gaseous
Effluents-Part 4: Specific requirements for radioactive iodine monitors.
IEC 60761-5 (2002) Equipment for Continuously Monitoring Radioactivity in Gaseous
Effluents-Part 5: Specific requirements for tritium monitors.
19
20
21
1.2
Special word usage
Throughout this standard, three verbs have been used to indicate the degree of rigor
intended for each specific criterion. The word shall denotes a requirement, the word
should denotes a recommendation, and the word may denotes permission.
22
2.0
18
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
Definitions
Acceptance Test: Evaluation or measurement of performance characteristics to verify
that certain stated specifications and contractual requirements are met as agreed
between the manufacturer and the purchaser/user. (ANSI N323A-1997; ANSI N42.332003)
Accredited Calibration Laboratory: A calibration laboratory that has been accredited
by an authoritative body (e.g., Health Physics Society, American Association of
Physicists in Medicine, National Institute of Science and Technology), with respect to its
qualifications to perform calibrations on the type of instruments covered by this
standard. (ANSI N323A-1997)
Accuracy: The degree of agreement of the observed value with the conventionally true
value of the quantity being measured. (ANSI N323A-1997; ANSI N42.33-2003)
Adjust: To alter the response by means of hardware or software control. (ANSI N323A1997; ANSI N42.33-2003)
Air Sampler: Device designed to trap the radioactive contamination contained in a
known volume of air passing through a filter or around an impaction surface in a pre-set
time interval. For example, continuous moving filter type, intermittent moving type, fixed
filter type. (IEC 50(394)
9
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
Air Monitoring Instrument: Radiation monitor for the continuous measurement of the
radioactive emission rate of the airborne materials. (IEC 50 (394)) Note: These are also
termed Continuous Air Monitors or CAMs.
Calibrate: (A) To adjust and or determine the response or reading of a device relative
to a series of conventionally true values for radiation sources. (B) To determine the
strength of a radiation source relative to a standard or conventionally true value (ANSI
N323A-1997).
Calibration: A set of operations that establishes the relationship between values
indicated by a measuring instrument or measuring system, and the conventionally true
values of the measurand under specified conditions. (ANSI N42.33-2003)
Check Source: A not necessarily calibrated radioactive source that is used to confirm
the continuing functionality of an instrument. (ANSI N42.33-2003)
Conventionally True Value (CTV) of a Quantity: The commonly accepted estimate of
the value of that quantity. This and the associated uncertainty will preferably be
determined by a national or transfer standard, or by a reference instrument that has
been calibrated against a national or transfer standard, or by a measurement quality
assurance (MQA) interaction with the National Institute of Standards and Technology
(NIST) or an accredited calibration laboratory. Refer to ANSI N42.22-1995 (reaffirmed
2002) and ANSI N42.23-1996 (Reaffirmed 2004).
Decade: A range of values for which the upper value is a power of ten above the lower
value (ANSI N323A-1997; ANSI N42.33-2003).
Detection Limits: Refer to Range.
Detection Threshold (Lower limit of detection): Value of the indication of the
measurement for which the relative random uncertainty equals +/-100% at the
probability of 95%. (IEC 394-40-20)
Detector: A device or component designed to produce a quantifiable response to
ionizing radiation normally measured electronically. (ANSI N323A-1997; ANSI N42.332003)
Effective Center: For a given set of irradiation conditions, the point within a detector
where the response is equivalent to that which would be produced if the entire detector
were located at the point. (ANSI N42.33-2003)
Energy Dependence: Variation in instrument response as a function of radiation
energy for a constant radiation type and exposure rate referenced to air. (ANSI N42.332003)
10
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
Functional Check: A frequently used qualitative check to determine that an instrument
is operational and capable of performing its intended function. Note: Such checks may
include, for example, a battery check, zero setting, or source response check. (ANSI
N42.33-2003)
Indicated Value: (A) A scale or decade reading. (B) The displayed value of the
readout. See also: reading (ANSI N323A-1997; ANSI N42.33-2003).
Instrument: A complete system consisting of one or more assemblies designed to
quantify one or more characteristics of ionizing radiation or radioactive material. (ANSI
N323A-1997, ANSI N42.33-2003)
Keep-alive Source: A radioactive source, typically with a disintegration rate of a few
Bq, mounted on or near the detector to assure that the detector is operational at all
times by providing a continuous positive signal.
Monitoring: Real time measurement of radioactivity or radiation level. (ANSI N42.332003)
Overload Response (or Over-range Response): The response of an instrument
when exposed to radiation intensities greater than the upper measurement limit. (ANSI
N323A-1997; ANSI N42.33-2003)
Performance Test: An evaluation of the performance of an instrument in response to a
given influence quantity. (ANSI N323A-1997; ANSI N42.33-2003)
Personal air sampler: Air sampler worn on the person (e.g., lapel air sampler).
Portable Air Monitoring Instrument: A readily transportable instrument that meets
the definition of “air monitoring instrument,” and that is not considered to be “installed.”
Range: All values lying between the detection threshold and the upper measurement
limit. (ANSI N323A-1997, ANSI N42.33-2003)
Reading: The indicated value of the readout. (ANSI N323A-1997, ANSI N42.33-2003)
Readout: The portion of the instrument that provides a visual display of the instrument
response. (ANSI N323A-1997, ANSI N42.33-2003)
Response: Ratio, under specified conditions, of the value of quantity measured by the
equipment or assembly under test and the conventionally true value of this quantity.
(IEC 50(394))
Sampling: The process of collecting a sample from a volume of air.
Standard (instrument or source): (A) National standard – a standard determined by a
11
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
nationally recognized competent authority to serve as the basis for assigning values to
other standards of the quantity concerned. In the U.S., this is an instrument, source, or
other system or device maintained and promulgated by the National Institute of
Standards and Technology (NIST). (ANSI N42.33-2004) (B) Primary standard – a
standard that is designated or widely acknowledged as having the highest metrological
qualities and whose value is accepted without reference to other standards of the same
quantity. (ISO 1993; ANSI N42.33-2004) (C) Secondary Standard - a standard whose
value is assigned by comparison with a primary standard of the same quantity. (ISO
1993; ANSI N42.33-2004) (D) Reference Standard - a standard, generally having the
highest metrological quality available at a given location or in a given organization, from
which measurements made there are derived. (ISO 1993; ANSI N42.33-2004) (E)
Transfer Standard: A standard used as an intermediary to compare standards. (Note:
If the intermediary is not a standard, the term transfer device should be used. (ISO
1993) (F) Working Standard - a standard that is used routinely to calibrate or check
material measures, measuring instruments, or reference materials. A working standard
is usually traceable to NIST. (ISO 1993; ANSI N42.33-2003)
Test: A procedure whereby the instrument, circuit, or component is evaluated. (ANSI
N42.33-2004)
35
3.0 Program Elements
36
37
38
39
40
41
42
43
44
45
Type Testing: Initial test of two or more production instruments made to a specific
design to show that the design meets defined specifications. (ANSI N42.33-2004)
Uncertainty: The estimated bounds of the deviation from the conventionally true value,
generally expressed as a percent of the mean, and ordinarily taken as the square root
of the sum of the square of the two components (ANSI N323A-1997; ANSI N42.332003):
a) Those components that are evaluated by statistical means; and
b) Those components that are evaluated by other means. (see NIST Special
Publication 1297).
Upper Measurement Limit: The upper measurement limit is the maximum
level at which the instrument meets the required accuracy. (ANSI N323A1997; ANSI N42.33-2003)
The test and calibration elements of a comprehensive air monitoring instrument
program include tests that are performed once (typically on new instruments), such as
type testing and acceptance testing, as well as elements that are performed repeatedly,
such as functional checks, periodic maintenance and recalibration, and performance
testing. Each of these elements is discussed further in the following sections.
Appendix A summarizes a life-cycle approach to the development and application of
instrumentation. It lists additional development and prototype steps that may be
performed when developing and applying an instrument. These additional tests are not
12
1
2
3
4
discussed further in this standard but are included in Appendix A to provide a
comprehensive outline of a complete program for instrument development, testing, and
application.
5
3.1
Type testing
The accuracy of operational measurements can be affected by instrument stability and
by variation in influence quantities such as radiation energy and radiation direction,
temperature, pressure, humidity, and electromagnetic and electrostatic conditions. The
instrument performance characteristics, as defined by the type testing, will indicate how
accurately the instrument can be expected to measure the airborne radioactive material
under variable operating conditions.
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
Type testing shall be performed for each instrument model. Type testing shall be
performed on two or more production units to fully characterize the performance and
limitations of the instrument. Users shall match operational requirements with the
conditions of type testing that demonstrate an instrument is capable of meeting those
requirements. The results of type testing shall be documented.
Type testing shall be repeated:
1. When modifications are made to the instrument design that may affect instrument
performance, or
2. When changes occur in operational conditions that may affect instrument
performance
3.2
Acceptance testing
Acceptance testing shall be performed on each new instrument before initial use. For
large numbers of instruments, it shall be considered acceptable to establish statistical
criteria for testing and accepting instruments.
Acceptance testing should evaluate each instrument against specific characteristics
identified during the type testing as critical or indicative of overall instrument
performance. The purpose of acceptance testing is to demonstrate that the instrument
meets certain stated specifications and/or contractual requirements.
An acceptance test should consist of:
a) physical inspection,
b) general operations tests, and
c) radiological response tests.
The acceptance testing protocol should be agreed upon between the manufacturer and
the purchaser/user. The results of acceptance testing should be documented.
13
1
3.2.1 Physical inspection
2
3
4
5
6
7
8
A physical inspection identifies any physical abnormalities that may effect instrument
operation (for example, broken parts; improperly operating moving parts; loose or
missing screws; unsecured circuit boards; loose wires, connectors, or components;
loose or misaligned knobs, and misaligned calibration potentiometer access holes).
Physical inspections shall be performed on each instrument.
9
3.2.2 General operations test
10
11
12
13
14
15
A general operations test is a determination of non-radiological operating functions (for
example, checking battery condition, verifying mechanical zero, testing the meter zero
potentiometer, or checking visual display elements, if applicable). General operations
tests shall be performed on each instrument.
16
3.2.3 Radiological response tests
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
When greater than 40 instruments are involved, the radiological response tests should
be performed on a random selection of 10% of the instrument batch. When fewer than
40 instruments are involved, no fewer than 4 instruments should be tested, or 100% of
the batch if fewer than 4 instruments. The instruments should be tested under the
range of anticipated operating conditions. If one instrument in a sample of a large
quantity fails the test, an additional 10% should be tested. An additional failure would
result in the need to test the entire batch.
Radiological response characteristics that may be tested during acceptance testing
include detection efficiency, energy dependence, overload response, measurement
range, and stability.
These tests shall be performed on calibrated instruments.
32
3.3
33
34
35
Each instrument shall be calibrated prior to deployment. Initial calibration shall be
performed as described in Sections 4, 5 and 6.
36
3.4
37
38
39
40
41
Initial calibration
Functional checks
Functional checks shall be performed in the field prior to putting the instrument into
service and periodically during use to ensure:
(1) that the instrument is operating properly, and
14
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
(2) that the instrument alarms on an appropriate level of detected radioactivity.
The frequency of functional checks should be commensurate with the overall reliability
of the instrument, the degree to which the instrument monitors itself for failures, and the
hazard potential of the material to be monitored.
The manufacturer should provide an appropriate method to verify:
a.
b.
c.
d.
Proper audible and visual alarms,
Detector operation using a radioactive source or natural background,
Flow rate, and
Calibration.
If the functional check includes the use of a check source to verify radiological
response, the radiological response of the instrument should be within ±20% of the
reference response. The reference response should be obtained for each instrument
when exposed to a source in a constant and reproducible manner, either at the time
that the instrument is received in the field or before its first use. Where it is shown that
readings for a specific model of instrument fall within ±20%, it shall be permissible to
obtain generic reference responses for a given source. Alpha monitoring instruments
shall use a long-lived alpha emitter such as thorium or uranium for functional checks.
Functional checks shall be documented. Appropriate data from the functional checks
should be recorded and reviewed for adverse trends indicating the need for changes in
component replacement frequency, component design, recalibration frequency, or
facility-related interfaces.
3.5
Maintenance and recalibration
Maintenance and recalibration shall be performed periodically on all instruments to
assure that the instruments continue to meet the required accuracy for field
measurements. This is required even when the functional check requirements are met.
Recalibration frequency shall be documented.
The frequency of maintenance and recalibration should typically be set at one year and
shall be based on the following:
 Documented functional check procedures and results,
 Documented maintenance histories including failure rates for the same model
instruments in the same environment,
 Manufacturer specifications,
 Design specifications, and
 Performance history.
If the performance history of the instrument(s) indicates that the instrument is
sufficiently reliable (i.e., recalibration as-found data is within tolerance > 95% of the
15
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
time), less frequent recalibrations may be performed, provided the recalibration
frequency does not exceed 3 years. An example motivation for less frequent
recalibrations is if the monitor is not routinely accessible (i.e. during periods of reactor
operation).
16
3.5.1 Maintenance
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
Where instruments are subjected to extreme operational conditions, hard usage, or
corrosive environments, more frequent maintenance and recalibration should be
scheduled. As found readings shall be documented to demonstrate that few (e.g., less
than 15%) instruments are out-of-calibration with the selected recalibration frequency.
If large numbers (e.g., more than 15%) of instruments are out-of-calibration at the end
of the selected recalibration interval, the maintenance and recalibration frequency
should be increased. NCSL RP-1, Establishment and Adjustment of Calibration
Intervals, provides additional information on a statistically based, technically defensible,
method of determining the appropriate recalibration intervals.
Maintenance shall be performed using components that affect performance, electronic
and/or electrical and mechanical, which are at least equivalent to those specified by the
manufacturer. If the manufacturer does not provide maintenance instructions, a
maintenance instruction should be written and approved by persons in the organization
responsible for performing the maintenance and/or recalibration. Where the
manufacturer does provide written instructions, the organization responsible for
maintenance and/or recalibration may document acceptance of the instructions.
Repairs made using unapproved instructions or components that may affect instrument
performance constitute an instrument modification, and shall render invalid any type
tests made on the instrument model as applied to the specific instrument. Modified
instruments shall have their performance tested and documented prior to issuance for
field use. If the user can document that the modifications will not affect the instrument
performance, additional testing is not required.
3.5.2 Recalibration
Recalibration is different from a functional check or simple evaluation with a check
source. It includes a pre-calibration check followed by adjustment and recalibration.
Recalibration shall be scheduled after any maintenance or repair that can affect
instrument performance. For this requirement, adjusting the air flow rate, adjusting
alarm set points, or changing batteries or filters is not normally considered
maintenance.
During the pre-calibration check the instrument shall be tested to ensure that certain
electronic operating requirements, specified by the manufacturer, are met. This
ensures that the instrument is in proper working order prior to recalibration.
16
1
2
3
Recalibration shall be performed as described in Sections 4, 5 and 6.
4
3.6
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
Performance tests
Over the life of an instrument, the instruments should be periodically performance
tested to verify that component aging and/or replacement components have not
affected the ability of the instrument to meet operational requirements. For large
numbers of instruments, it shall be considered acceptable to establish statistical criteria
for performance testing.
Whenever modifications are made to the instruments that affect their response,
performance tests shall be conducted to evaluate the response characteristic that may
have been impacted. Modifications include altering the instrument firmware or software
or replacing instrument parts with spare parts other than those recommended by the
manufacturer.
Whenever changes that may affect instrument performance occur in the operational
environment, performance tests shall be conducted to evaluate how the instrument
performance may have been impacted. Such tests may be conducted as an extension
of the original type testing.
All or a subset of the performance tests may be conducted as part of the maintenance
and recalibration procedures. Performance testing may include range, detection
efficiency, linearity, detection limit, and response to overload conditions.
Performance tests shall be documented and documentation of any changes in the
performance characteristics of the instrument shall be provided to the user.
4.0
Air flow rate calibrations
All flow meters shall be calibrated against devices that are traceable to NIST or another
National/International Standards Laboratory. Before calibrating the air flow-rate sensor,
the air in-leakage should be verified to be less than 5% (ANSI N13.1-1999) of the
nominal flow rate. Flow meters shall be calibrated with a filter or device installed that
produces a pressure drop equivalent to the pressure drop of the filter type to be used
during normal operations.
During calibration, the standard flow meter should be upstream of, and inline with, the
filter so that it measures air going through the filter. This will prevent measuring air that
has entered the system from leakage paths around or after the filter.
Rotameters and mass flow meters shall be calibrated at the average flow rate
anticipated during use, and should also be calibrated at one point above and one point
17
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
below the anticipated sampling flow rate (e.g., 75% and 125% of the anticipated
sampling flow rate).
21
5.0
22
23
24
25
This section provides general instructions applicable to all air monitors. Additional
instructions for specific types of air monitors are provided in the Section 6.
26
5.1
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
Critical venturi flow meters shall be calibrated with a sufficient pressure differential
across the meter to ensure that the velocity at the throat of the meter is sonic. The
absolute pressure at the entrance of the critical flow meter should be within 
the average absolute pressure anticipated at that location. The temperature at the
entrance of the critical flow meter during calibration should be within 5C of the
average temperature anticipated at the entrance to the critical flow meter during
sampling.
Air flow meter readings should be within ±15% of the conventionally true value of the
actual flow rate. Note that if the instrument uses a mass flow meter, appropriate
corrections for ambient pressure and temperature are required. The air flow reading
shall correspond to actual volumetric flow, not standard flow at standard temperature
and pressure.
Additional information on calculating uncertainty for flow meter calibrations is provided
in Appendix B.
Instrument calibrations
Pre-calibration inspection
Prior to calibrating the instrument, inspect the instrument for obvious damage. After
determining that there is no obvious damage, establish the following conditions prior to
exposing the air monitor to a radioactive source for adjustment and calibration:
1. Instrument contamination levels shall be low enough that contamination does not
interfere with calibration. The detector and sample chamber should be as free of
radioactive contamination as possible to prevent cross contaminating calibration
sources.
2. If the instrument is zeroed as part of normal operation, the meter or display shall be
adjusted to zero or the point specified by the manufacturer using the adjustment or
adjustments provided.
3. The batteries and/or power supply shall comply with the instrument manufacturer's
specification.
4. The instrument shall be turned on and allowed to stabilize as appropriate.
18
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
5.2
As-found readings
Prior to adjusting the instrument, as-found readings shall be taken and recorded. The
as-found readings should include:
1. determining the system background and verifying that it is within the userestablished limits, and
2. determining the response of the instrument to a source of known radioactivity and
verifying that it is within the user established limits.
If separate calibration controls are used for each range or decade, as-found readings
shall be taken on each range or decade. If only one calibration control is used for all
ranges or decades, only one as-found reading needs to be recorded.
As-found readings shall also be recorded for the flow sensor.
5.3
Electronic calibration
Electronic calibration shall be performed to verify meter/display linearity. Linearity shall
be checked at several points within the range of the instrument. If the instrument has a
single calibration adjustment for the meter or display, electronic calibration need be
performed at only two points; one near the lower limit of detection and one near the
upper limit of measurement. If multiple calibration adjustments are provided, two points
within the range of each adjustment shall be checked.
The signal used to perform electronic calibration should mimic the pertinent
characteristics of the detector signal (e.g., pulse width, pulse height).
In addition to verifying/adjusting linearity, the electronic calibration may include adjusting
pulse width, obtaining high voltage (HV) plateaus, and setting detector operating
voltage, amplifier gain, thresholds, and windows.
34
5.4
Meters and chart recorders
35
36
37
38
39
Both local and remote monitor recorders, if supplied, should be verified to provide
readouts within ±10% of the meter indication. This verification should be made at a
minimum of three points covering the range of the instrument.
40
5.4.1 Analog linear readout instruments
41
42
Analog, linear readout instruments with a single calibration control for all scales shall be
19
1
2
3
4
5
6
7
8
9
10
11
12
13
14
adjusted either at the point recommended by the manufacturer or at a point within the
normal range of use. Instruments with calibration controls for each scale shall be
adjusted on each scale.
15
5.4.2 Analog logarithmic readout instruments
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
If instruments have ranges that are not calibrated, a limited calibration tag that notes
the limits of calibration shall be attached to the instrument.
After adjustment the response of the instrument shall be checked near the end points of
each scale (approximately 20% and 80% of full scale).
Instrument readings shall be within 15% of the conventionally true values (CTV) for the
lower point and 10% of the CTV for the upper point. Readings within 20% shall be
acceptable if a calibration chart or graph is prepared and provided with the instruments.
Analog, logarithmic readout instruments that commonly have a single readout scale
spanning several decades, normally have two or more adjustments. The instrument
shall be adjusted for each scale according to the site specifications or the
manufacturer’s specifications. Alternatively, it shall be permissible to calibrate at points
of importance to the user if the impact on accuracy is justified and documented.
If instruments have ranges or decades that are not calibrated, a limited calibration tag
that notes the limits of calibration shall be attached to the instrument. The same
principles should be applied to microprocessor-controlled instruments.
After adjustment, calibration shall be checked at a minimum of one point on each
decade. Instrument readings shall have a maximum deviation from the CTV of no more
than 10% of the full decade value. If the display is not marked at intervals
corresponding to 10% of the point (i.e., a display calibrated at a decade mark
frequently does not have graduations to 10% in the next decade), it is permissible to
use the spacing on the low side of the calibration point to estimate the calibration limit
on the high side of the calibration point.
5.4.3 Digital readout instruments
Digital readout instruments shall be calibrated as in 5.4.1.
If the instrument is designed to auto scale, the calibration point should be selected far
enough from the auto scaling point that auto scaling will not affect the reading.
The instrument should be cycled through a complete test of all display segments or
indicators, either electronically or radiologically.
20
1
2
3
If the instrument has both digital and analog readouts, both readouts shall be
calibrated.
4
5.4.4 Microprocessor-based instruments
5
6
7
8
9
10
If the instrument uses multiple algorithms or if the influence of calibration factors varies
over the range of application, the calibration or test points selected for the instrument
shall include at least one point within each algorithm and transition points between
algorithms.
11
5.5
12
13
14
15
16
17
Verify that background count rate due to electronic noise, instrument contamination, or
external radiation sources is within limits. If the instrument uses a filter, background
readings should be taken with the filter in place. Air flow should be off to prevent the
accumulation of radon progeny on the filter during the performance of this test.
18
5.6
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
Background response verification
Calibration of background subtraction
Proper functioning of background subtraction capability shall be verified. If the
instrument corrects for interference from radon progeny collected on a filter, a filter shall
be installed and air flow shall be on during the performance of this test.
If the instrument responded to interfering external radiation fields during type testing,
calibration adjustments for background subtraction should be performed in appropriate
interfering external radiation fields equivalent to the radiation fields expected during
use.
Calibration of the background subtraction capability may be performed electronically
using pulse generators or other electronic calibration devices.
The manufacturer’s recommendation should be followed for calibration of monitors that
have background subtract capabilities.
35
5.7
Alarm circuit verification
36
37
38
39
40
41
42
Calibration shall include verification of radiation and failure alarms, alarm delay time,
and alarm relays. Operability of audible and visual alarms may be verified electronically
or by using a radioactive source to actuate the alarm. Alarm set points shall be set and
verified during the electronic calibration and should actuate at ±10% of the expected
value. If monitors have adjustable alarm set points that may be adjusted during use,
this alarm setting mechanism shall also be verified to function properly.
21
1
2
3
4
5
6
7
8
9
5.8
Radiological calibration
Real-time air monitoring instruments should be calibrated with radioactive sources that
exhibit the radiation type and energies of the airborne radioactivity to be monitored.
Table 2 identifies a selection of sources that should be used for calibration.
Table 2 - Reference sources for calibration of continuous air monitors
Radiation type
Gamma
Alpha
Beta
10
11
12
13
14
15
Reference source
137
Cs filtered to remove ≥ 90% of
the betas and associated 137Ba
x-rays
241
Am, 230Th, 239Pu, NatU
90
Sr-90Y in equilibrium, filtered
through 50 mg/cm2 of material
with atomic number (Z) ≤ 13
Efficiency measurements should be performed in accordance with the manufacturer's
recommendations. Radiation sources of sufficient strength to provide a count rate of at
least one-quarter full scale should be used to determine efficiency.
17
6.0 Additional calibration requirements for particulate and noble
gas monitors
18
19
20
In addition to the requirements of Section 5.0, the following requirements apply to the
calibration of specific types of air monitoring instruments.
21
6.1
16
22
23
24
25
26
27
28
29
30
31
32
33
Alpha and beta particulate monitors
Calibration of alpha and beta air monitoring instruments using single or multi channel
analyzers shall include steps to verify or adjust linearity, pulse width, gains, windows,
and thresholds. These steps may be performed electronically using a pulse generator
or similar electronic calibration devices that provide periodic and variable amplitude
signals or they may be performed using radioactive materials. These steps shall be
performed prior to performing the radiological calibration.
Energy calibration of alpha monitors that use multi-channel analyzers should be based
on radon progeny peaks (radon progeny collected on a filter) to ensure that the solid
source characteristics don’t influence the energy calibration. Response to a plated
source should then be adjusted (as appropriate) to compensate for differences between
22
1
2
the response to a plated source and the response to a filter source.
3
6.2
4
5
6
7
8
9
10
11
12
13
14
15
Keep-alive sources
The absence of significant quantities of alpha emitting radon progeny may cause
detector low count alarms on an alpha continuous air monitor (CAM) placed
downstream of a High Efficiency Particulate Air (HEPA) filter. Keep-alive sources may
be used to prevent spurious alarms. For example, a small amount of thorium (a few
tenths of a Bq generating around 2 to 3 cpm, depending on instrument efficiency) may
be mounted on or near the detector.
When this technique is used the outside of the instrument should be labeled to indicate
that a source is present, and appropriate adjustments to the background subtraction
circuit should be made.
16
6.3
17
18
19
20
Although radon is a natural radioactive noble gas, its measurement is not included in
this standard. Note, however, that the presence of radon, or its decay products, may
interfere with the measurement of other radioactive gases.
21
6.3.1 Pre-calibration
22
23
24
25
Pre-calibration should include verifying that pre-filters or absorbants are in place to
remove particulates and radioiodines.
26
6.3.2 Electronic calibration
27
28
29
30
31
32
33
34
Noble gas monitors
To prevent the need for several sources of different activities, the measuring assembly
alone may be tested by injection of an appropriate electronic signal at the normal
detector input.
Electronic set up of the noble gas monitor should be consistent with the electronic set
up used during initial calibration.
35
6.3.3 Radiological calibration
36
37
38
39
40
41
Gaseous sources should be used for calibration. If solid sources are used, type testing
shall establish the response of the instrument to the gas of concern. In addition,
response of the unit to solid sources shall be established by cross calibration against
gaseous sources during type testing.
23
1
2
3
4
5
6
7
8
Gaseous sources may consist of bottles of pressurized air or gas containing the
radioactive gas with which the unit is to be tested. This may be the radioactive gas for
which the unit is designed or some other radioactive gas of interest.
9
6.3.3.1
Solid sources may be used in lieu of radioactive gas sources for subsequent calibration.
Such sources shall be of physical form appropriate for the unit being calibrated, to allow
location of the source relative to the detector to be accurately fixed and repeatable.
Initial Calibration with Gaseous Sources
10
11
12
13
14
15
16
17
18
19
20
21
Gaseous sources shall be used for the type testing of noble gas monitors. The type
testing may also include the initial calibration. The gas used should be the gas intended
to be monitored. Depending on the application and the type of detector in use with the
instrument, calibration with a gas not to be monitored, may be used provided that
corrections for the difference in energies are made. The calibration gas used shall be
traceable to NIST or another National/International Standards Laboratory except in the
case where the calibration laboratory has the capability to perform a traceable analysis.
Tests using the calibration gas shall be corrected as needed for temperature, pressure,
radioactive decay, time and chamber or system volume. The calibration will determine
the value of the measured gas vs. the instruments response.
22
6.3.3.2
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
Calibration with solid sources
Calibration of noble gas monitors with solid sources is allowed only after an initial
calibration with gaseous sources during type testing has determined the instrumental
efficiency. Solid sources used in the calibration of gaseous monitors shall be traceable
to NIST or another National/International Standards Laboratory. The sources shall be
presented to the detector under test in a documented and reproducible geometry. Initial
testing with solid source(s) shall be done to determine the correlation of the solid source
efficiencies to those determined during the initial calibration performed with gas. An
Initial evaluation with solid source(s) shall determine the value of the measured
source(s) vs. the instruments response. The efficiency determined for the solid source
shall be compared with the value determined from the gaseous source calibration to
produce the response factor to be used in all subsequent calibrations utilizing the solid
sources. All subsequent calibrations using the solid sources shall, at a minimum, test
two points of the operating range of the detector. One point within the first two decades
above the background level of the system and the other within the upper two decades
of the system shall be tested to insure system linearity.
All solid source calibrations shall be corrected for radioactive decay and use the solid
source / gaseous source ratio for the determination of the noble gas detector efficiency.
7.0
Calibration for special conditions
24
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
An instrument should be used within the operating limits and for the types of radiation
for which it was designed and type tested.
If the instrument will be used under environmental conditions that are outside the
operating limits of the instrument as identified by the manufacturer or during type
testing, a calibration for special conditions shall be performed.
If an instrument will be used to measure activity from a source other than for which it
was designed (i.e., using a tritium monitor to measure krypton gas or using a
transuranic alpha particulate monitor to measure radon progeny), a special calibration
shall be performed with a radioactive source similar in geometry and physical properties
to the source the instrument will be used to monitor.
The calibration for special conditions needs to be performed only once and a correction
factor established to use in later calibrations, unless:
1. The instrument is modified or physically altered, or
2. The normal calibration suggests a change or deterioration has occurred in
instrument performance, or
3. The special conditions change.
Instruments calibrated for special conditions shall be labeled with the conditions under
which the calibration is valid.
26
8.0
Facilities, conditions, and equipment
27
8.1
Calibration facility
28
29
30
The photon radiation background at the calibration facility should be low, known, and
stable and shall be accounted for and documented during calibration.
31
8.2
32
33
34
35
36
37
38
39
40
Test and calibration conditions
Testing and calibration of instruments should be performed under a group of controlled
conditions called standard test conditions, Table 3. This is recommended to ensure that
performance variations are eliminated in the calibrations and that performance
characteristics can be related to the calibrated instrument. Temperature, relative
humidity, and atmospheric pressure shall be noted at the time of instrument
calibrations.
Table 3 – Standard test conditions
25
Influence quantities
Acceptable range for standard test conditions
Warm up time
> 1 minute or manufacturer’s specification
Relative humidity
Ambient ±10%, not to exceed 75%
Ambient temperature
20C to 24C
Atmospheric pressure
70 to 101.3 kPa
Line voltage(a)
Nominal ± 1%
Frequency(a)
60 Hz ± 0.5 Hz
Background ambient photon
radiation (external)
2.5% of full scale of the range or decade under
test, but nominally should not exceed 0.5 μGy/h
(50 μrad/h), referenced to air.
Non-ionizing electromagnetic field
of external origin
Less than 50% of the lowest value that causes
interference
Magnetic induction of external
origin
Less than twice the induction due to the earth's
magnetic field
Residual radionuclide
contamination
Contamination shall be low and should be less
than limits for total activity listed in NRC Reg Guide
1.86.
(a) Applicable only to AC powered instruments.
1
2
3
4
5
6
7
8
9
10
11
12
13
8.3
Calibration standards
Instruments shall be calibrated with appropriate standards of known hierarchy derived
from national standards. Working standards, when used, shall be compared at least
annually to a higher level standard. Refer to ANSI N42.22 and ANSI N42.23.
Calibrations of the reference radiation fields or sources shall be obtained in one of the
following ways:
1. Comparison of the radiation field from a user's reference source with the radiation
field from a national or secondary standard source in the same geometry, using a
"transfer instrument" with a reproducibility of ±2 percent. The transfer calibration
26
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
shall use a calibration curve for the transfer instrument taken with the national or
secondary source over a range that covers both the national or secondary source
measurement and the user source measurement. (Such a curve reduces to a single
point if the transfer calibration procedure is such that the transfer instrument
readings are identical for both measurements.)
2. Calibration of a user's transfer instrument with a national or secondary standard
source, followed by evaluation of user's reference source with the same transfer
instrument. The transfer instrument shall have a reproducibility of ±2 percent and
the procedure shall use a calibration curve as in Section 8.3 (1).
3. Where no national or secondary standard exists (as in the case of specific energies,
unusual sources, or unusual source configurations), by establishment of a standard
source or instrument with documented empirical and theoretical output or response
characteristics.
A calibration source or set of sources should be of sufficient activity to reach full scale
of any instrument to be calibrated. If the source is a radionuclide, the half-life should be
long, preferably greater than several years to minimize corrections and uncertainties.
Calibration sources should be from Table 2 and shall comply with applicable standards.
The activity or surface emission rate of calibration sources shall be known with an
accuracy of +/-10%.
Working standards shall be used only for routine calibrations or testing of instruments.
The source configuration and geometry should be comparable to the air sample
collector to detector configuration, and should be reproducible. The activity distribution
of working standards should mimic the collection media activity distribution. A
correction factor may be used to compensate for differences between the deposition
characteristics of the working standard and the collection media. Working standards
used to calibrate air monitoring instrumentation should approximate as closely as
possible the actual media and isotopic energy being monitored.
The energy spectrum for all calibration sources should be known and if absorbers,
collimators or high scatter geometries are employed, the effect shall be known on the
calibration accuracy relative to that from a low-scatter geometry. When shutters or
other source control devices are used, the effect of transit time on dose shall be known,
accounted for and documented.
The activity of calibration sources should be characterized and documented. Such
considerations as charged particle equilibrium, scattered or unwanted radiations from
the source, and ambient background radiation should be accounted for during
standardization of the calibration source.
8.4
Maintenance of standards
45
27
1
2
3
4
5
6
7
Calibration standards shall be maintained through an ongoing measurement quality
assurance (MQA) program with NIST or an accredited calibration laboratory. Direct
comparison with national or secondary standards shall be undertaken when the MQA
program indicates an uncertainty in calibration greater than 4%. Quality control
procedures shall be in place to assure that the working and reference standards at a
facility maintain a known relationship.
8
8.5
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
Check sources
Check sources should provide radiation of the same type or types as provided by those
sources used in instrument calibration. However, check sources may provide radiation
different from that used for calibration if:
(1)
The source-to-detector instrument geometry is well understood and easily
reproduced, or
(2)
The instrument response to this radiation is well understood and is not critically
dependent on instrument adjustment (e.g., the use of a photon source to check
instruments calibrated to detect to beta radiation may be acceptable; the use of a
photon source to check a detector utilizing a BF3 tube response to neutrons is not
acceptable). A reproducible source-to-detector geometry shall be established and
used for all functional check measurements.
Check sources do not need to be traceable to NIST. To avoid functional check failures
due to degradation of the check source; however, check sources should be compared
periodically with standards.
27
9.0
Documentation
28
9.1
Facility documentation
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
A comprehensive and readily available record system shall be maintained by the
calibration facility. All records of data shall identify the individual who collected the data
on which the record is based. The facility records shall include at least the following:
1. The entire results of type testing
2. An inventory of all standards and calibration equipment.
3. A full history and calibration data, including certificates, for all standards and
applicable calibration equipment.
4. All procedures used for providing calibration services.
5. Routine quality control records.
28
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
6. The results of all performance testing.
7. Records of instrumentation for which a calibration service was provided, and the
date that the calibration was performed including a description, sufficient for
identification of every item. The record system shall also include or reference a
detailed report for that specific calibration and the name of the person performing
that calibration.
8. Copies of all calibration reports issued and/or electronic records.
9. Information essential to the analysis and reconstruction of the calibration of a
specific item of instrumentation. We see little difference between this and 4.
10. Records detailing the education, experience, and training of all operating staff and
supervisory personnel.
11. A documented analysis of calibration uncertainty.
9.2
Instrument Documentation
Records shall be maintained of all calibration, maintenance, repair, and modification
data for each instrument. The record shall be dated and shall identify the individual
performing the work. The record shall be filed with previous records on the same
instrument and shall be readily retrievable. Additional information on calibration records
including recommended retention periods can be found in ANSI/HPS N13.6-1999.
Each instrument shall be labeled with the following information:
1.
2.
3.
4.
5.
Date of most recent calibration,
Initials or other specific identifying mark of the calibrator,
Date that calibration is again required,
Special-use or limited calibration label (if applicable), and
Serial number of instrument or other unique identification number used by the
facility to identify a specific instrument.
39
10.0 References
40
41
ANSI N1.1-1976, Glossary of Terms in Nuclear Science and Technology.1
1
ANSI publications are available from the Sales Department, American National
Standards Institute, 11 West 42nd Street, 13th Floor, New York, NY 10036, USA.
29
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
ANSI N13.1-1999, Sampling and Monitoring Releases of Airborne Radioactive
Substances from the Stacks and Ducts of Nuclear Facilities.
ANSI N13.2-1969 (Reaffirmed 1982), Administrative Practices in Radiation Monitoring.
ANSI N320-1979, Performance Specifications for Reactor Emergency Radiological
Monitoring Instrumentation.
ANSI N323A-1997, American National Standard Radiation Protection Instrumentation
Test and Calibration, Portable Survey Instruments.
ANSI N42.17A-1989 (Reaffirmed 1994), Performance Specifications for Health Physics
Instrumentation - Portable Instrumentation for Use in Normal Environmental Conditions.
ANSI N42.17B-1989, Performance Specifications for Health Physics Instrumentation –
Occupational Airborne Radioactivity Monitoring Instrumentation.
ANSI N42.18-1974 (Reaffirmed 1991), Specification and Performance of On-Site
Instrumentation for Continuously Monitoring Radioactivity in Effluents.
ANSI N42.30-2000, Performance Specifications for Tritium Monitors.
ANSI N42.33-2004, American National Standard for Portable Radiation Detection
Instrumentation for Homeland Security.
ANSI/HPS N13.6-1999, Practice for Occupational Radiation Exposure Records.2
ASTM C859-92B, Standard Terminology Relating to Nuclear Materials.3
ICRU Report 60-1998, Fundamental Quantities and Units for Ionizing Radiation.4
ICRU Report 39-1985, Determination of Dose Equivalents Resulting from External
Radiation Sources.
2
ANSI/HPS publications are available from the Health Physics Society, 1313 Dolley
Madison Blvd, Suite 402, McLean, VA, 22101, USA.
3
ASTM publications are available from ASTM International, 100 Barr Harbor Drive,
West Conshohocken, PA, 19428-2959 USA
4
ICRU publications are available from the National Council on Radiation Protection and
Measurements, 7910 Woodmont Avenue, Suite 800, Bethesda, MD 20814, USA.
30
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
ISO 25-1990, General requirements for the competence of calibration and testing
laboratories5
IEC 50(394)-1995, International Electrotechnical Vocabulary, Chapter 394, Nuclear
Instrumentation, Instruments
ISO 4037-1 - 1996, X and Gamma Reference Radiations for Calibrating Dosimeters
and Dose Rate Meters and for Determining Their Response as a Function of Photon
Energy – Radiation Characteristics and Production Methods.
ISO 8769-1988, Reference sources for the calibration of surface contamination
monitors – beta-emitters (maximum beta energy greater than 0.15 MeV) and alphaemitters.
NBS Special Publication 250-1989, Calibration Services User's Guide. 6
NCSL RP-1-1996, National Conference of Standards Laboratories, Establishment and
Adjustment of Calibration Intervals.7
NIST-2001, NIST Handbook 150, 2001 Edition, Procedures and General Requirements
NIST Special Publication 1297-1993, Guideline for Evaluating and Expressing the
Uncertainty of NIST Measurement Results
NRC Regulatory Guide 1.86, Termination of Operating Licenses for Nuclear Plants. 8
NRC Regulatory Guide 8.25 (US NRC 1992)
ISO 1993, ISO Guide 99, International Vocabulary of Basic and General Terms in
Metrology (VIM), International Organization for Standardization, 1993
5
ISO publications are available from the ISO Central Secretariat, Case Postale 56, 1
rue de Varembé, CH-1211, Genève 20, Switzerland/Suisse. ISO publications are also
available in the United States from the Sales Department, American National Standards
Institute, 11 West 42nd Street, 13th Floor, New York, NY 10036, USA.
6
NBS and NIST publications are available from the Superintendent of Documents, U.S.
Government Printing Office, Washington, D.C. 20402.
7
NCSL publications are available from the National Conference of Standards
Laboratories, 1800 30th St. Suite 305B, Boulder, CO, 80301, USA.
8
NRC publications are available at www.nrc.gov.
31
1
2
3
4
Appendix A - Summary of a Life-Cycle Approach to Instrument Development and
Application
Life-Cycle Step
Mission
Evaluation a
Research and
Development a
Prototype Testing a
Type Testing b
Production Control
Testing a
Training a
Acceptance
Testing b
Initial Calibration b
Functional
Checks b
Purpose
To identify
mission
objectives and
conditions,
measurement
objectives,
performance
requirements,
and candidate
technologies
To aid in the
development of
a prototype that
is likely to meet
specifications
To demonstrate
that the design
of the
instrument is
likely to meet
certain
specifications
To demonstrate
that the design
of the
instrument as
manufactured
meets certain
specifications.
To control
production,
avoid defects,
and confirm
instrument
compliance with
specifications,
To ensure
proper execution
of all tests and
requirements
To demonstrate
compliance with
selected
specifications.
To establish a
traceable
calibration
relevant to
expected
conditions of
use.
To provide
indications that
the instrument is
operational.
Timing or frequency
Prior to initiation of the
development process, and
continuously throughout
the life cycle
Units to be tested
None (this step is
conceptual in nature)
Specifications
Definition of the
specifications that
will be evaluated in
the subsequent lifecycle steps
Responsibility
Responsible
officials, with input
from research,
development,
manufacturing, and
users
As needed
Individual
components and
assemblies.
As selected by the
manufacturer; or
requested by the
purchaser/user.
Manufacturer
As needed prior to start
of production
One or more
prototype units
As selected by the
manufacturer; or
requested by the
purchaser/user.
Manufacturer
(generally);
purchaser/user
(occasionally)
A minimum of once prior
to full production.
Two or more initial
production units.
All specifications
from the relevant
standard, or as
agreed upon between
manufacturer and
purchaser/user.
Generally the
manufacturer.
Occasionally the
purchaser/user
Depending on acceptable
failure rate agreed upon
between manufacturer
and purchaser/user.
As determined by the
manufacturer or as
agreed upon between
manufacturer and
purchaser/user.
As selected by
manufacturer, or
requested by the
purchaser/user.
Manufacturer
As appropriate to support
proper execution of each
step of the instrument life
cycle
After the units are
received and prior to their
initial use.
As needed to train
individuals to
properly carry out
life-cycle steps
As agreed upon
between
manufacturer and
purchaser/user.
Each unit.
As appropriate for
the instrument and
its use
All life cycle
participants
As selected by the
purchaser/user.
Purchaser/user.
Selected instrument
parameters and
responses.
Designated
calibration staff of
the users’
organization (or
selected vendor).
Each unit.
As appropriate for
the instrument being
used.
User.
Prior to initial use.
Before each use and
periodically during use.
32
Life-Cycle Step
Operational
Experience a
Purpose
Timing or frequency
Units to be tested
Specifications
Responsibility
To provide
During each operation
A representative
Selected instrument
User.
evidence of any
and during periodic
number of units
parameters and
operational
reviews of experience
responses
deficiencies
To provide
At a frequency, such as
Each unit.
As appropriate for
Designated
Maintenance and
preventive
annually, based on the
the instrument.
maintenance staff of
Recalibration b
maintenance,
design and reliability
the user’s
make necessary
history of the instrument.
organization (or
repairs, and
selected vendor).
reestablish a
traceable
calibration.
To verify that
As appropriate based on
A representative
Selected
As arranged by the
Periodic
the instrument
experience and
number of units.
specifications from
purchaser.
Performance
continues to
anticipated modes of
the type test.
Testing b
meet relevant
failure.
specifications.
a.
This step is included in the table to provide a complete overview of a comprehensive program for instrument development,
testing, and application. This step is not discussed further in this standard.
b. This step is part of a comprehensive instrument test and calibration program. This step is discussed in further detail
elsewhere in this standard.
(This table was adapted from Hoover, M.D. and Cox, M., A Life-Cycle Approach for Development and Use of Emergency Response and Health
Protection Instrumentation, in Public Protection from Nuclear, Chemical, and Biological Terrorism, Brodsky, A., Johnson, R.H. Jr., and Goans,
R.E., Eds., Medical Physics Publishing, Madison, WI, 2004.)
1
33
1
APPENDIX B - Flow Rate Meter Calibrations
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
The goal of the flow rate meter calibration is to help ensure that the uncertainty in the
measurement of the total volume of air sampled is  10%.
The following approach, described in U.S. Nuclear Regulatory Commission (NRC)
Guide 8.25 (US NRC 1992) can be employed to calculate the total uncertainty in the
volume of air (Uv):
2
2
2 1/2
UV = (Us + Uc + Ut )
where: Us is the uncertainty (dimensionless) in reading the flow rate meter scale. This
can be estimated by dividing one-half the value of the smallest scale division by
the indicated flow rate.
Uc is the error (dimensionless) associated with determining the calibration factor,
i.e., correcting the indicated flow. As an approximation, the error associated
with the calibration instrument may be used.
Ut is the error (dimensionless) associated with measurement of the sampling
time.
Note that Us can be multiplied by a fluctuation term Fs to account for fluctuation in the
steady-state reading of the meter. The fluctuation term is set at 1 for a meter whose
readings do not fluctuate. If there are fluctuations, the fluctuation term is taken to be
the average observed number of scale units above and below the mean indicated
value.
34
Download