GCE Physics

advertisement
1
PH5 Magnetism, Nuclei & Options
General
PH5 is intended as a terminal unit, though the regulations do not require A level Physics to be
cashed in at the same time as PH5 is taken. The 1¾-hour examination paper has 3 sections:
Section A
This is a 60-mark section based upon the core content of PH5.1 – PH5.5. It is designed to
be answered in about 60 minutes.
Section B
This section carries 20 marks and contains a series of questions relating to the Case Study.
It is designed to be answered in about 20 minutes.
Section C
This section consists of 5 questions, 1 on each of the Optional Topics. The questions each
carry 20 marks and are designed to be answered in about 20 minutes. Candidates will
answer 1 question only.
SECTION A – Core content
In line with the other units of this specification, PH5 is designed to require a teaching time of
approximately 60 hours, of which ¾ should be devoted to the study of the compulsory core
content.
PH5.1 – Capacitance
This topic is part of the national core and follows on from the electrostatic fields section of
PH4. It also draws on energy and electrical circuits concepts. The equation for the decay of a
capacitor is of the same form as that for radioactive decay. It is a good topic for the
introduction of semi-log graphs for the determination of the time constant of the decay and
hence the capacitance.
There are many examination questions in past PH4 papers and the topic is well covered in
most A-level physics textbooks.
PH5.2 – B-fields
This section deals with the concept and definition of magnetic fields, their effect on moving
charges (in wires and in free space), their production and application in particle accelerators.
This traditional topic is well covered in A-level text books and there are many examination
questions to be found in past PH5 papers.
PH5.3 – Electromagnetic Induction
In addition to magnetic flux (linkage) and the laws of electromagnetic induction, this topic is
linked via rotating coil generators to the basic concepts of alternating current electricity –
frequency, period, peak values, r.m.s. values. The relationship between peak and r.m.s. values
for a sinusoidally varying quantity and the use of r.m.s. current and voltage in power
calculations are explored. Candidates will be expected to have used oscilloscopes to measure
voltages and currents [by the p.d. across a resistor] and frequencies. This traditional topic is
well covered in A-level text books and there are many examination questions to be found in
past PH5 papers.
2
PH5.4 – Radioactivity and Radioisotopes
This core topic is largely unchanged from the previous specification and is well covered in Alevel text books. Candidates will be expected to handle logarithm and exponential functions.
Calculations on decay can be expressed via the exponential function e   t or 2 x , where x is
the number of half-lives. This topic area lends itself to synoptic questions which combine
concepts of relative atomic mass and the mole with the decay equations to calculate the
activity of a given mass of material of known decay constant. Past PH5 papers contain many
examination questions of an appropriate level.
PH5.5 – Nuclear Energy
This core topic is largely unchanged from previous specifications. As with radioactivity, the
mole concept will be used in estimating the energy release from macroscopic quantities of
reacting materials – the electron volt is also a concept from AS which is of use here.
Conservation of mass/energy s introduced, using E = mc2. This concept can be applied
generally and not only in particle interactions – e.g. calculate the Sun’s power output, and
hence the Solar Constant, given that its mass loss per second is 4 million tonnes.
A useful concept is the energy equivalence of 1 u. 1 u = 1.6604  1027 kg. The energy of this
mass is 1.6604  1027  c2 = 1491  10-10 J = 931 MeV.
The use of neutrino energy in neutrino detectors is possible. The common reaction used in
detectors is:
37
17
Cl  ν e 
38
17
Ar 
1
0
e
The masses of the particles are:
Cl: 36.96590 u
e: 0 [at least negligible]
Ar: 36.96677
e: 0.000548 u
The gain in mass in the interaction Δm = 36.96677 + 0.0005  36.96590 = 0.00142 u  1.32
MeV. This means that only neutrinos with a kinetic energy of more than 1.32 MeV can cause
this interaction. The mean neutrino energy produced by the first step in the proton-proton
chain is only 0.26 MeV, so most of these are not detected.
SECTION B – the Case Study
Centres with candidates for PH5 will receive, in February of the relevant year, multiple
copies of a printed passage based upon a physics topic of contemporary interest. The topic
will be chosen to relate to previously studied areas of the AS and A2 specification. It will be
assumed that candidates have covered the whole of the AS specification and PH4 at least.
The passage should be given to candidates for study and teachers are encouraged to discuss
its contents with them, drawing their attention to the relevant areas of the specification and
considering the sorts of questions which could arise from its contents. Candidates will be
provided with a clean version of this passage in the PH5 examination and will not be
permitted to take notes into the exam.
3
Previous PH6 examinations contain similar passages with the difference that, in the legacy
specification, candidates had not previously seen the passage and were expected to read it in
the examination. Nevertheless, these passages form a good resource for introducing this
section of the paper. Because candidates will be expected to have studied the passage prior to
the examination, no allowance for reading is built into the duration of PH5, which is 105
minutes.
SECTION C – Options
There are 5 optional topics:
A
B
C
D
E
Further Electromagnetism and Alternating Currents
Revolutions in Physics
Materials
Biological Measurement and Medical Imaging
Energy Matters
Each topic is designed to be studied in approximately 15 hours of teaching time. They could
all be taught at the end of PH5. They fit in with the rest of the specification in different ways,
which suggests that different teaching strategies are appropriate:
Option A follows on immediately from the electromagnetism and A.C. material in PH5. The
filters section relates also to the potential divider ideas in PH1.
The Electromagnetic Revolution aspect of Option B, which will be the setting for questions
for the first 3 years, relates in the early stages to the optics material in PH2, the electrostatics
in PH4 and the electromagnetism in PH5. There is a strong case, if this option is to be
offered, for incorporating its ideas throughout the teaching of the rest of the course,
Option C, materials, consists of ideas which were previously in the compulsory specification
and are now optional. There are few strong links with other sections of the specification.
Option D, Biological Measurement and Medical Imaging, has links to PH2 and PH5.
Option E, Energy Matters, links to PH1, PH2 and PH4 and so a possible approach would be
introduce the content throughout the course.
Guidance notes on each of the options follows:
4
Unit PH5 Option A – Further Electromagnetism and Alternating Currents
The majority of this option unit is taken from the legacy specification where it was
compulsory content. This historical material divides into 3 parts:
1. Mutual induction and its application to transformers. This treatment is largely qualitative,
except for the treatment of the ideal transformer in terms of turns ratio and the equality of
input and output powers.
2. Self inductance and inductors.
3. Phasor analysis of series RC, RL, LC and RCL circuits.
This material is adequately covered in text books and PH5 papers. Questions on transformers
also appear in past GCSE Physics papers. It is not the current intention to produce teacher
guidance notes on these aspects of Option A.
Statements (p) – (r) deal with the sharpness of a resonance curve (Q- factor) and the
application of the RC potential divider to high pass and low pass filter circuits. These topics
are well dealt with in Electronics text books and past GCE Electronics papers [ET4] of WJEC
but have not been dealt with in previous WJEC GCE Physics specifications.
The Quality factor (Q) of a resonant circuit
The quality (Q) factor of a LCR circuit is related to the sharpness of the resonance curve. A
high Q factor gives a sharp resonance curve while a low Q factor gives a broad resonance
curve (see the diagram below with Q = 8 and Q = 2).
The main component in determining the Q factor of the circuit is the resistance of the circuit
because it is the resistance that dissipates energy away from the circuit. This is similar to
pushing a swing back and forth – if there is a lot of friction taking energy away from the
swing it’s difficult to achieve a high amplitude and ‘sharp’ resonance. The easiest way to
define the Q factor is as follows
Q
r.m.s. pd across inductor at resonance
r.m.s. pd across resistor at resonance
As the capacitor and inductor have equal reactance at resonance, the Q factor can also be
written:
5
Q
r.m.s. pd across capacitor at resonance
r.m.s. pd across resistor at resonance
These definitions lead to the equations
Q
I 0 L

IR
0 L
R
I
and also Q 
0C

IR
1
0 CR
If we also incorporate the expression for the resonant frequency, 0 
Q
0 L
R
 1
L


1
LC 


R
R
1
, then
LC
L
C
So we have three expressions for the Q factor.
Q
0 L

R
1
0 CR

1
L
R
C
Note that, in the expressions for the Q factor, we can eliminate L, C and ω0 but we cannot
eliminate R – it is in all 3 expressions. Note also that the Q factor is a ratio and it has no units.
Now consider this circuit:
10 V
~
10 nF
10 mH
10 
These values for R, C, L make our arithmetic reasonably easy. They give us the following
figures:
0 
1

LC
1
2
10  10
8

1
10
10
 105 s -1
and
Q
0 L
R

105  102
10
 100
We can also calculate the current flowing at resonance because the whole of the supply p.d. is
across the resistor at resonance (p.d.s across the inductor and capacitance are equal and
opposite, so cancel).
I
V
R

10
10
1 A
All seems nice and straight forward until we look at the p.d. across the capacitor or inductor.
6
VL  I 0 L  1 105  102  1000 V   VC 
How can we have 1000 V across the inductor (and capacitor) when the supply voltage is only
10 V? There is no simple answer to this question but a better understanding can be drawn
from considering another type of resonance. Again, consider a swing with very little friction.
You only need to provide a small push regularly in order to obtain a large amplitude – you
might only be pushing the swing for a distance of 30 cm but the amplitude of oscillation
could easily be 2 m.
Understanding how a series LCR circuit can be used to select frequencies
We know now that LCR circuits with high Q factors can increase the p.d. in an a.c. circuit.
This can be used in the design of a simple radio. The circuit below can be used as the
detection part of a simple radio. It consists of an antenna (long wire), inductor, variable
capacitor and earth connection.
A good application of synoptic physics:
antenna
Remember that radio waves are electro-magnetic waves and
have oscillating electric (and magnetic) fields. These oscillating
fields will cause electrons to move in the metal antenna. The
moving electrons will give us an alternating current and an
alternating p.d. (due to the resistance of the antenna).
earth
If you look at the loop in the above circuit, you’ll notice that there is no resistor. It is an LC
circuit without the R. Why is this? Remember that we want a high Q factor and one of the
ways that this is achieved is to keep the resistance low. Does this mean that the resistance in
the LC loop is zero? Obviously the resistance cannot be zero because the connecting wires
aren’t made of superconductors. But the main source of resistance in the LC loop is the
inductor. Remember that an inductor is a long wire wound into a coil. In order to make a
large number of loops we need a thin wire and this increases the resistance of the inductor (a
bit of a Catch 22 situation).
A simplified way of analysing the performance of the detecting circuit above is to consider it
as follows:
~
VOUT
i.e. we have a series LCR circuit and the voltage across the variable capacitor is the output
voltage. Note also that we have redrawn the inductor as a resistor and inductor in series
because of the inherent resistance of the wires of the inductor.
7
If we have resonance in the LCR circuit we know (from the definition of the Q factor) that the
p.d. across the capacitor will be Q times the supply p.d. Hence, we can amplify the input p.d.
by a factor of Q. Also, because of the shape of the resonance curve we only amplify the
frequencies around the resonance frequency, so we have selectivity.
So why do we use a variable capacitor? This is because we can vary the resonance frequency
by varying the capacitance. We obtain the resonance frequency from the equation below.
f0 
0
1

2 2 LC
So, the above simple circuit does three things:
1. Amplification - it amplifies our signal
2. Tuning - it can tune to a particular resonance frequency (by changing C)
3. Selectivity - it amplifies only those frequencies around the resonance frequency.
If you would like to see this tuning circuit in operation there is a reasonably priced kit
available from Maplins - N51FL crystal radio £5.99. Alternatively, it is possible to design
and build your own radio using instructions available from many internet sites e.g.
http://www.midnightscience.com/cigar.html, http://journeytoforever.org/edu_radio.html,
http://www.electronics-tutorials.com/receivers/crystal-radio-set.htm .
~
Example
This circuit is used in a simple radio.
(i) Calculate the Q factor when C = 6 pF
and when C = 600 pF.
(ii) Calculate the range of frequencies
to which the circuit can tuned.
10 
0.15 mH
6  600 pF
8
Using CR circuits as Low Pass and High Pass Filters
NOTE:
One thing you must beware is that you cannot simply add the rms p.d.s across the resistor and
capacitor in either of these a.c. filter circuits.
Remember always, for r.m.s. p.d.s ,
VIN  VC2  VR 2
and that (in general)
VIN  VC  VR
Low pass filter:
R1
Compare with
R
~ Vin
C
R2
Vout
The easiest way to explain how the above circuit behaves as a low pass filter is to compare it
with a voltage divider.
In the circuit on the right, the supply voltage is shared between the two resistors. In the low
pass filter, on the left, the voltage is divided between the capacitor and the resistor.
Remember that the reactance of the capacitor is given by: X C 
1
C
From the above equation, at low frequencies XC will be very large. So at low frequencies we
have a voltage divider with a very large “resistance” in the R2 position. This means that
nearly all the supply voltage will be across the capacitor at low frequencies.
At high frequencies XC will be very small. So at high frequencies we have a voltage divider
with a very low “resistance” in the R2 position. This means that nearly all the supply voltage
will be across the resistor at low frequencies i.e. there will be a very low p.d. across the
capacitor.
If we were to draw a graph of Vout/Vin against frequency we would get:
9
Low pass filter output
1.0
Vout/Vin
0.8
0.6
0.4
0.2
0 0
10
10
2
10
4
10
6
Frequency/Hz
Note that Vout/Vin is usually called the gain and that it starts at 1 and drops to zero (this is
because Vout = Vin at very low frequencies and Vout = 0 at very high frequencies).
Example
1kΩ
10V ac supply
1nF
Vout
~
Vout
1. Calculate the frequency when the rms p.d.
across the resistor is equal to the rms p.d. across
the capacitor.
2. Calculate the rms p.d. across both the resistor
and the capacitor at the frequency of Q1.
Answers
1
Equating the p.d.s across the capacitor and resistor we get:
IX C  IR . Cancelling I gives us: X C  R
1
1
1
 R and rearranging we get  
But X C 
, hence
.
C
C
CR

1
1


 159 kHz .
Using   2 f , we get: f 
9
2 2 CR 2 10 1000
Beware:
There are 3 pitfalls to avoid if you want to obtain the correct answer even after you’ve
1
obtained the equation f 
.
2 CR
 First, you must remember that kΩ means 1000 Ω.
 Second, you must remember that nF means 10-9 F.
 Third (but this only applies if you have an EXP button on your calculator and if
you’re too lazy to do the powers of 10 in your head!), when putting 10-9 in your
calculator you cannot type 10 exp -9 because this is the same as 1010-9. You must
type (and this might seem strange until you think about it carefully) 1 exp -9 because
this is 110-9.
10
2
There are many ways of obtaining the correct answer
V
e.g.using Z  X C2  R2 and I  etc.
Z
but it is probably more direct and simple to do as follows:
Remember that VS2  VC 2  VR 2 and that VC  VR from question 1.
VS2
V
10
and hence VC  S 
 7  07 V
2
2
2
So the correct answer is that the p.d. across both the capacitor and the resistor is 7.07 V.
So VS2  2VC 2 ,  VC 2 
Beware:
Do not fall into the trap of saying that both rms p.d.s must be 5V so that they add up to 10V.
Although this sort of argument applies to instantaneous p.d.s it is completely wrong for
obtaining rms p.d.s because the p.d. across the capacitor is out of phase with the p.d. across
the resistor.
High pass filter:
In the low pass filter of the previous section we noted that when the p.d. was low across the
capacitor the p.d. was high across the resistor. If we now swap our capacitor and resistor, the
output p.d. will be the p.d. across the resistor instead of the capacitor (see below left). In this
circuit we will have a high output where we previously had a low output and a low output
where we previously had a high output. See the graph at the bottom of the page and compare
it with the previous low pass filter graph. The graph at the bottom of this page is
characteristic of a high pass filter.
C
Compare with
R1
~ Vin
R
Vout
R2
Again, the easiest way to explain how the above circuit behaves as a high pass filter is to
compare it with the voltage divider (on the right).
In the circuit on the right, the supply voltage is shared between the two resistors. In the high
pass filter, on the left, the voltage is divided between the capacitor and the resistor.
1
C
From the above equation, at low frequencies XC will be very large. So at low frequencies we
have a voltage divider with a very large “resistance” in the R1 position. This means that the
output voltage across the resistor at low frequencies will be close to zero.
Again, remember that the reactance of the capacitor is given by: X C 
11
At high frequencies XC will be very small. So at high frequencies we have a voltage divider
with a very low “resistance” in the R1 position. This means that nearly all the supply voltage
will be across the resistor at high frequencies.
If we were to draw a graph of Vout/Vin against frequency we would now get:
High pass filter output
1.0
Vout / Vin
0.8
0.6
0.4
0.2
0 0
10
10
2
10
4
10
6
Frequency / Hz
Note:
Filters are usually drawn in the following manner:
R
VIN
VOUT
C
0V
This makes it easier to draw higher order filters (i.e. one filter feeding into another to provide
more filtering). This notation has not been used here so that students can compare the circuit
more easily with a potential divider. However, the above notation may well be used in an
examination.
12
Unit PH5 Option B – Revolutions in Physics
OPTION B: REVOLUTIONS IN PHYICS
ELECTROMAGNETISM AND SPACE-TIME
1.
Lifetime of the Electromagnetism and Space-time material
When a History of Physics option was proposed, two periods of revolutionary change immediately
suggested themselves for study: the century of Kepler, Galileo and Newton, and the century of
Young, Faraday and Maxwell. Rather more interest was expressed in the second of these, and only the
Electromagnetism and Space-time revolution will be examined in 2010, 2011 and 2012. After the new
A-Level has been running for a year or two, teachers will be consulted on whether or not a change
should be made for examinations in 2013 and beyond.
2.
Content
One of the most exciting things in Physics is to discover relationships between phenomena which are
seemingly very different in nature. What happened in electromagnetism in the nineteenth century is a
wonderful example. In the year 1800 there were only the vaguest indications that magnetism had
anything to do with moving electric charges, and no evidence at all that light had anything to do with
electricity or magnetism. By 1900 magnetism and electricity had been firmly linked, and light had
been shown to be an electromagnetic wave.
The seemingly obvious need for electromagnetic waves to have a propagation medium (the ether)
created problems. These were resolved in a very radical way by Einstein’s Special Theory of
Relativity.
The structure of the course is shown in a little more detail in the diagram below. The first main
‘block’ deals with events leading to the acceptance of the wave theory of light, starting with a careful
look at Thomas Young’s description of his two slits experiment. Electromagnetism is the subject of
the next main block, starting with Ørsted’s discovery of the magnetic effect of a current, and
considering at some length the subsequent work of Ampère and Faraday.
Maxwell arrived at the conclusion that light was an electromagnetic wave using what we would now
call a mechanical model of electric and magnetic fields. How he made the synthesis is looked at in
some detail, as are the beautifully simple confirmatory experiments of Hertz. The Michelson-Morley
experiment is then outlined, as are responses to its failure to yield the expected evidence for the ether.
Finally there is a small taste of Special Relativity theory (a simple treatment of time dilation)
A brief survey of light, electricity and magnetism before
1800
(NEWTON, HUYGENS, GILBERT, GALVANI, VOLTA)
is followed by a more detailed study of…..
YOUNG
FRESNEL
ØRSTED
AMPÈRE
FARADAY
synthesis
MAXWELL
HERTZ
MICHELSON
EINSTEIN
13
3.
Serving Suggestions
All the material to be tested in the examination is contained in the 34 sides of WJEC notes, which are
available in electronic form from the Physics section of the WJEC website <link to be inserted> or as
hard copy from the WJEC subject officer.
The notes contain many self-test questions and could be used by a student for self-study. They are
also peppered with links to websites which help to bring the basic material of the option alive and
make it easier to learn. The sites often contain pictures and diagrams.
The first half of the course (Young, Ørsted, Ampère, Faraday) deals largely with concepts in light and
electromagnetism which are key parts of the non-optional A-level specification, but comes at them
from a different angle, adding ‘human interest’, and (obviously) a historical perspective. The result
should be reinforcement. A possible teaching strategy is to integrate the material of the first part of
this option with the normal teaching of the relevant topics. The second half of the material might lend
itself to self-study with lessons on specific topics, such as time dilation.
4.
Extracts from writings of Young, Faraday, Maxwell, Hertz, Einstein
The extracts contained in the WJEC notes are short but they do give the student something
approaching direct contact with great physicists of the past. They are supported by explanatory notes
and self-test questions to help with understanding. In the examination, part of the Option B question
might present the student with a snippet from one of the extracts and ask him or her to explain certain
points, or to put the extract in its historical context.
Those who associate studying history with the enforced learning of dates need not have too many
fears about this option. Placing discoveries in the right half decade will suffice.
5.
Books
Two thinnish and very readable books which provide good support are…
Michael Faraday and the Royal Institution: by John Meurig Thomas (ISBN 0-7503-0145-7).
Relativity and its Roots: by Banesh Hoffmann (ISBN 0-486-40676-8).
Chapter 4 tells pretty much the same story as this course, but, as the book’s title makes clear,
Hoffmann has a special agenda, and his emphases are different.
Examination questions are restricted, however, to the material in the WJEC notes (though the student
will be assumed to have tackled the embedded self-test questions – which occasionally call for him or
her to find facts elsewhere).
14
Unit PH5 Option C – Materials
General
The optional unit is designed to provide students with a sound understanding of the behaviour
of materials. Emphasis is placed on understanding the relationship between the physical and
mechanical properties of metals, glasses ceramics, polymers and composites and the
microstructure of these materials.
The unit also touches upon the use of advanced materials such as superalloys and carbon
fibre reinforced plastics and their applications.
It is hoped that the unit will provide students who have an interest in pursuing a career in
engineering or material science with an increased technological awareness of the field of
materials and to alert them to the possibilities that the future holds in this rapidly advancing
and increasingly important field.
Much of the unit is covered in sufficient depth in standard A-level texts. This coverage is
dealt with in Table 1. The text books referred to are detailed below the table. Further useful
material is to be found in the following:
1. Cooke (B) and Seng (D) (1989). Physics of Materials for A-level students (2nd ed).
Leeds. University of Leeds.
2. Advanced Physics project for independent learning (APPIL) – unit: Behaviour of
matter. John Murray.
3. Easterling (K) (1990). Tomorrow’s materials. London Bourne Press.
4. Gordon (J.E.) 1978. Structures, or living things don’t fall down. Great Britain.
Penguin Books.
5. Gordon (J.E.) 1976. The New Science of Strong materials or why you don’t fall
through the floor. Great Britain. Penguin Books.
Useful web pages. The links are active in the electronic version of these guidance notes and
every attempt will be made to ensure their current validity.
1) The Macrogalleria- a cyberwordland of polymer fun.
www.pslc.ws/mactest/index.htm
2) Stress-Strain curves.
www.shodor.org/~jingersoll/weave/tutorial/tutorial.html
3) www.s-cool.co.uk
4) www.antonine-education.co.uk
(Follow links to AS Physics and then to Module 3, Topic 6 – elastic properties of
solid materials).
5) www.schoolphysics.co.uk
15
Table 1.
Selected
Book
Reference
Specification
Reference
a
Adams & Allday
Duncan
10.3
p. 17 – 19, 24
b
c
d
e
3.8 (p.60 – 61)
10.1
f
10.2 (p. 440 –
441)
10.2 (p. 440 – p. 28 – 29
441)
10.6 (p. 448)
p. 31 – 34
g*
h*
i*
j
k*
l
m*
n
o*
p*
q
r
10.11 (p. 458
459)
10.6 (p. 448)
10.2 (p. 441)
10.10 (p. 456
457)
10.7 (p. 450)
10.8 (p. 452
453)
10.8 (p. 452
453)
10.9 (p. 454
455)
10.10 (p. 456
457)
p. 29
p. 30
p. 39
Advanced
Physics
for Muncaster
You
p. 288 – 289,
291
p. 282
p. 284 – 285
p. 286
p. 287
9.10 (p. 147 –
151)
11.2 (p. 182)
p. 183 – 184
11.5 (p. 184 –
185)
11.12, 11.13 (p.
192 – 193)
– p. 35 – 36
p. 290
p. 33
p. 28 – 29
– p. 34 – 35
p. 289, 290
11.10,
11.11
(p.191)
11.3 (p. 193)
p. 190
p. 290
– p. 41
p. 291
p. 291
– p. 41
p. 291
– p. 37
p. 293
11.8, 11.9 (p.
189 – 190)
p. 189
–
p. 38
p. 294
* Further guidance given in the Teacher Guidance Notes or in associated documents.
Texts referred to in the Table:
1. Duncan G.T. (1987) Physics, A textbook for Advanced Level Students (2nd ed).
London. John Murray Ltd..
2. Adams (S) and Allday (J) (2000). Advanced Physics. Oxford. O. U. Press.
3. Johnson (K) et al (2000). Advanced Physics for You. U. K. Nelson Thornes.
4. Muncaster (R) (1993). A-level Physics. Cheltenham. Stanley Thornes.
16
Detailed guidance
Specification references (a) – (f), (j). (l), (n), (q) and (r) are treated in sufficient detail in
standard A level text books to obviate the need for guidance in this document.
Elastic and Plastic Strain
The process of deformation of ductile materials, including the movement of edge
dislocations, is treated at the molecular level in the WJEC document:
The plastic behaviour of ductile metals.
SUPERALLOYS
Statement (i) draws upon statements (h) and (j). Candidates should have an understanding of
the effects of dislocations at the molecular level, and the strengthening and stiffening of
materials by the introduction of dislocation barriers such as foreign atoms, other dislocations
and grain boundaries (specification statement (h)). Candidates should also be able to describe
failure mechanisms in ductile materials and have an understanding of creep and fatigue
(specification statement (j)).
Introduction
Aircraft jet engines are required to operate within extreme conditions of temperature and
pressure. Jet engine turbine blades rotate at a typical speed of 10,000 rpm for long periods in
an environment of combustion products at working temperatures of 1250ºC (though the inlet
temperatures of high performance engines can exceed 1650ºC); non aviation gas turbines
operate at approximately 1500˚C. The blades must be able to withstand impact and erosion
from debris drawn in with the air stream. In addition, different parts of the blade may be at
different temperatures and they will be subjected to large and rapid temperature changes
when the engine is started up and turned off.
The following is a list of the properties required of the material from which the blades are
made:
 Creep Resistance
Centripetal forces acting on the blade at high rotational speeds provide a considerable
load along the turbine blade axis. Over prolonged periods of time this can cause creep. It
becomes increasingly pronounced as temperature increases. Creep could cause a turbine
blade to deform sufficiently that it might touch the engine casing.

Corrosion Resistance
Iron corrodes to form rust. At high temperatures, the presence of carbon dioxide, water
vapour and other products of the combustion of fuel constitute a highly corrosive
environment.

Toughness
The blades must resist impact with debris passing through the engine. In addition,
stresses generated by expansion and contraction, between different parts of the blade at
different temperatures, must not give rise to cracking.
17

Mechanical and Thermal Fatigue Resistance
Variations of gas pressure and temperature on different parts of a blade and mechanical
vibrations may generate cyclical stresses which can cause failure due to fatigue.

Metallurgical Stability
The mechanical properties of metals can be modified by heat treatment. Blade materials
must be resistant to such changes and the microstructure must remain stable at high
temperatures.

Density
The density must be low to keep engine weight as low as possible.
The separate document “Superalloys” http://www.wjec.co.uk/uploads/publications/15254.doc
which can be found in the GCE Physics section of the WJEC website, gives details of how
engineers have worked to produce single crystal turbine blades which satisfy these design
criteria.
HEAT TREATEMENT OF METALS
Strength and hardness are two mechanical properties of a solid metal which are affected by
crystal grain size. The smaller the grains, the stronger a material is- fracture is more difficult
with small grains because there are more grain boundaries and dislocations (moving within
one grain) have difficulty passing into adjoining grains. The atomic planes of adjoining grains
are in different directions so fractures tend to be halted at grain boundaries. The more
boundaries there are, the stronger the material is. The mechanical properties of metals can be
controlled by the following common treatments:

Work hardening.
This is a process which makes a metal stronger. The metal is worked or deformed (by
hammering or repeated bending) when cold to make it stronger and harder. The effect of
working the metal is to increase the number of dislocations, so increasing its strength. The
effects of work-hardening can be felt by bending the wire of a steel coat hanger
backwards and forwards until it snaps.

Quench hardening.
Suggested experiment: Heat one end of a 20cm (approx) length of steel wire (held with
tongs) in a Bunsen flame until it becomes cherry red in colour- about 800ºC. Then plunge
the hot end of the wire quickly into cold water. When the rest of the wire has cooled try to
bend the quenched end. What do you notice?
Rapid cooling ‘freezes’ a particular grain structure into the metal. The higher the
quenching temperature the smaller the grains and the harder and more brittle the resulting
metal.

Annealing.
Suggested experiment: Use the same sample of wire as above and heat the other end of
until it is red hot, and keep it at red heat for about 15 seconds. Withdraw it from the heat
very slowly so that it cools gradually. When cool, try to bend the annealed end. What do
you notice? This experiment can be carried out using a length of copper instead of steel.
Slow cooling allows grains to grow larger, making the metal softer, more easily bent,
hammered or scratched.
18
Griffith cracks and brittle failure of amorphous solids
A. A. Griffith, investigating the breaking stress of glass in the 1920s, estimated that its value
should be about 1010 Pa. The glass under study had a breaking stress of only 108 Pa but he
found that very thin fibres had much higher breaking stresses, with fibres of diameter 10 -3
mm having a breaking stress of 3  109 Pa. Thus, the breaking stress varies with the diameter
of the glass rod and approaches the theoretical breaking stress as the diameter of the rod
decreased. The results obtained are sketched in the following graph.
6
Strength/
Arbitrary
units
1
10
20
Diameter/ μm
The glass fractures by a process known as brittle fracture. This is accelerated by the
presence of surface imperfections or cracks.
This is shown in the diagrams on the next page. The stress becomes concentrated around the
tips of a crack. Bonds near the crack will break, increasing the load on neighbouring bonds
which are still intact, causing them to break and the crack propagates rapidly [at
approximately the speed of sound in the glass.
Crack
Key: The pecked
lines like this are
called stress lines.
They represent the
way the tensile load
is transmitted along
the rod from bond to
bond. Notice how the
lines are
concentrated near the
tip of the crack.
19
Stress lines and stress concentrations can be photographed by making specimens out of
Perspex and stressing them between crossed polaroids. The pictures on the next page show
this. The picture on the left shows stress lines in a uniform bar which is stressed. The bar on
the right has a small crack half way up its left side, resulting in a high concentration of stress
around the tip.
This diagram represents the atoms and bonds around the tip of a crack:
The force in bond A will be large since it has to balance the
forces exerted on molecules X and Y from above and
below.
The top two lines are incomplete because of the crack, so
that the stress they carry is transferred to the line of atoms
below. The bond A at the bottom of the crack is therefore
carrying a much higher stress than the rest of the bonds.
The stress can exceed the breaking stress of the material
only in this region causing the bond to break, increasing the
size of the crack and also the stress concentration. The
crack will therefore propagate quickly through the material
causing it to fracture.
In the case of the glass fibres, surface cracks are caused
among other things by differential cooling at the surface and in the centre. The narrower the
thread, the more uniform the temperature and so the less significant are any cracks that form.
This makes the small diameter rods much stronger. The very narrowest glass threads [~ 1
m] approach the theoretical strength predicted by Griffith. For very narrow threads,
inducing cracks by simply touching the surface brings their tensile strength back to that of the
bulk glass.
This property of brittle materials is exploited by glaziers when “cutting” a piece of glass to
size by putting a scratch in it and then snapping it – similarly with tiles.
Since amorphous solids break by brittle fracture, they will be weak under tension, but under
compression they will be very strong as the stress will cause the cracks to close preventing
propagation. When amorphous solids such as brick are used for building, the structures
produced are strong provided the material is kept under compression.
20
Experiments to investigate the strength of glass.
1) Heat the end of a glass rod so that it softens enough to bend it into a hook shape to support
weights. When cooled, heat the centre of the rod until soft, then remove the rod from the
flame and quickly draw it out into a fibre. (An alternative method for producing fibres is to
hang a 100g mass from the hook, heat the centre of the glass and as the glass softens and
begins to fall, immediately withdraw the heat. The mass should be allowed to fall through a
distance of about 0.5m to 1.0m). Then clamp the straight end so the fibre hangs vertically.
Load the fibre with weights until it snaps. Note the final weight and, using a micrometer,
measure the diameter of the broken fibre. Collating class results would allow graphs of
breaking force or breaking stress versus diameter or cross-sectional area to be drawn.
2) Support a glass rod horizontally at either end. Load its centre with weights until it snaps.
Repeat the test with the glass marked with a glass cutter underneath at the middle. The cut
rod should snap more easily.
3) Heat a glass rod and pull out a fibre about 0.5 m long (as in experiment 1). When it is cool
bend it into an arc on the bench. Release it, run your fingers over the fibre and bend it again.
The fibre should now snap more easily.
THE BEHAVIOUR OF RUBBER & POLYTHENE
A polymer macromolecule consists of long chain molecules, each containing up to 105 atoms
and these chains are held together by cross-bonds, the structure being similar to that shown in
the diagram.
Diagram 1. A possible arrangement for three long-chain rubber molecules.
The behaviour of a polymer depends on the strength of the cross-bonds and examples of two
types of polymer are considered below.
(i) Rubber – an example of polymer with weak cross bonds. Natural rubber is a polymer
of the molecule iso-prene. It has weak van der Waals cross-bonds and only a few covalent
(strong) cross-bonds. Its behaviour under increasing stress is shown by the following graph.
B
stres
s
A
O
strai
n
21
Between points O and A, the deformation is elastic and from the slope of the graph it can be
seen that, after an initial stiff region, Young’s modulus is small. At stresses greater than A,
the deformation is still elastic, but the value of Young’s modulus is much greater. If suffi
iently stressed, the material breaks.
Initially, as the rubber is deformed, no bonds are extended; the long chain molecules are
straightened against their thermal motion (which tends to increase the amount of folding in
the molecular chains). The van der Waal’s bonds are responsible for the initial stiff region,
but once they are overcome, the rubber molecules unfold and the material can extend by
several times its original length. Because bonds are not being broken here, the additional
stress needed to do this is small. The structure of the material changes as shown in the
following diagram.
_______ molecule 1
- - - - - - molecule 2
Diagram 2. Stretched rubber
At point A, the molecules the sections of the molecules which are free to unwind are more or
less straight, therefore if any further extension of the material takes place, bonds are
stretched. This is far more difficult to do than straightening the molecules, therefore the
value of Young’s modulus increases at this point.
When the stress is removed the thermal motion of the chain molecules makes the polymer
return to its original dimensions. The value of Young’s modulus for such a polymer
increases with temperature, the opposite to the variation in crystalline and amorphous solids.
This is due to the fact that the chain molecules have to be straightened against their thermal
motion. As the temperature increases, the thermal motion increases the amount of folding, so
that the average end-to-end distance in an individual molecule decreases, with the result that
straightening the molecules becomes more difficult.
These polymers also show elastic hysteresis i.e.
the stress-strain curves for the loading and
unloading do not coincide. This is shown in
this graph:
B
stress
A
OAB is the stretching curve and BCO the
C
contracting. The strain for a given stress is
greater when unloading than for loading. The
strain
unloading strain can be considered to ‘lag
O
behind’ the loading strain.
The area under OAB represents the work done [i.e. the energy supplied] to cause stretching;
similarly the area under BCO represents the energy given up by rubber during contraction.
22
This closed curve is called a hysteresis loop; its area is the energy per unit volume converted
into internal energy [or, ‘lost as heat’ in common parlance] during one cycle. Thus when a
polymer is repeatedly stressed, its temperature increases. Rubber with a hysteresis loop of
small area is said to have resilience. This is an important property where the rubber
undergoes continual compression and relaxation as does a car tyre when it touches the road
as it rolls on. If the rubber used did not have a high resilience, there would be appreciable
loss of energy resulting in increased petrol consumption and lower maximum speed. Heat
build up could even lead to tyre disintegration.
Not for examination: Chemical engineers alter the properties of natural rubber by the process of
vulcanization, in which strong covalent bonds are deliberately introduced between the long
molecules. This has the effect of making the rubber stiffer and increasing its resilience. This very
stiff form of rubber is useful for applications which involve repeated deformation, e.g. car tyres.
Polymers exhibit a property called creep during which the chain segments slowly disentangle
under a constant stress as a consequence of the thermal motion of the chain segments.
On the release of stress, thermal motion restores the mixing, but slowly, since the segments
get in each other’s way during the shuttling process.
(ii)
Polythene- a semi crystalline polymer.
Polythene is a thermoplastic polymer which contains both crystalline and amorphous
regions. Since these regions co-exist within one structure, polythene can be described as a
semi-crystalline material.
In the crystalline regions polymer chains composed of covalent C-C bonds fold together
many times in an ordered arrangement as shown.
Force
Force
These lamellae, as they are called, (from the word laminate) form small, spherical grains
called spherulites. [Note: The crystalline regions of a thin film of high density polythene
can be observed using polarised light and a microscope]. Between these crystalline grains,
the polymer molecules are tangled together in an amorphous state, where little long range
order exists. The graph below shows how the stress varies with the strain for a strip of
high density polythene when it is stretched.
23
Between O and A, the strip can regain its original length as its behaviour is elastic. The
parallel parts within the lamellae crystal are held together by weak van-der-Waals bonds
and at low strains these bonds resist the applied stress.
From A to B a ‘neck’ forms in the strip as the tangled molecules in the amorphous
regions start to align with each other.
From B to C the width of the neck remains unchanged as the strip is extended. In this
region the van-der Waals forces between the lamellae are overcome and they begin to
unfold and become parallel to each other.
The strip is now said to be cold-drawn and is very strong along the axis of the applied
stress. Beyond C, the stress is resisted by the strong covalent bonds between the carbon
atoms within each polymer molecule.
Investigating force–extension curves for rubber and polythene.
In these experiments you will investigate how rubber and polythene behave under
tension. It is not intended that you should obtain accurate values for the mechanical
properties of these materials, but basic quantities such as the elongation at fracture and the
breaking strength may be determined from the force-extension or stress-strain curves.
(a) (i) Rubber Band (cross-linked polymer).
(1) Hang a (cut) rubber band of (approximate) cross-section 1mm by 2mm vertically from
a stand, boss and clamp (Hoffmann clips are useful here to suspend the rubber band).
The base of the stand should be secured using a G-clamp. Attach another Hoffmann
clamp to the other end of the rubber band and use this to hang a 50gram mass holder
from. Place a metre rule as close as possible to the mass holder. The extension may be
read using an optical pin attached to the base of the mass holder.
(2) Measure the length, width and thickness of the rubber when it is supporting the 50
gram holder. Try to avoid squashing the rubber with the micrometer screw gauge.
(3) Increase the mass in 50 gram steps. Depending on the thickness of the rubber, you
may need to change the smaller masses for a single 0.5 kg mass in order to exceed the
elastic limit. Continue to add 50 gram masses until the rubber band breaks.
24
(4) Plot the force extension curve and determine the Young Modulus from the linear
section.
(ii) Natural Rubber.
A similar experiment can be carried out with a natural rubber strip, (linear polymer chains
with little or no cross-linking) cut so its width is about 5 mm. follow the procedure as for
the rubber band, increasing the mass in 50 gram steps up to 500 gram. You will probably
need to cater for an extension greater than 1.0 m. From the force-extension curve you
should be able to identify two regions which are approximately linear. If this is possible,
calculate two values of E for natural rubber.
(b) (i) Low density Polythene.
(1) Cut a strip of polythene about 1 metre in length from a thin polythene bag. You may
need to fold the polythene several times in order to get measurable thickness using a
micrometer screw gauge. You can then calculate the thickness if a single sheet.
(2) Increase the mass, initially in 50 gram steps, measuring the extension for each mass
added. The polythene, at some stage, will increase in length without any further
increase in load (A to C in above graph). If you plot a stress-strain graph as you go
along it is possible to find an accurate value for this increase in extension. Using 5
gram and 10 gram masses will also help you identify the point where this region
begins. When this region ends (beyond C), larger loads will be needed to produce any
further extension.
(ii) High density polythene.
Cut a strip of high density polythene from the rings used to hold cans of drink
together (a good excuse here for buying beer!!). Mark the strip at two points along its
length and, using the above procedure, slowly stretch it as much as possible without
breaking it. After the neck forms, observe how it lengthens without becoming
narrower. The strip is now cold-drawn and you can measure the sample’s breaking
stress.
The behaviour of polythene when
it is stressed.
At the point B, the chains are straight and any additional strain is due once more to the
stretching of bonds, thus the behaviour shown beyond this point is once more elastic. At C, the
material breaks.
25
Unit PH5 Option D – Biological Measurement and Medical Imaging
General
The optional unit deals with several different medical imaging and measurement techniques:
 X-rays including production, absorption, use in diagnosis and therapy and CAT
scans.
 Ultrasound including acoustic impedance and Doppler techniques.
 The principles of Magnetic Resonance Imaging including comparison with X-rays
and Ultrasound for imaging.
 ECG including the interpretation of the cardiogram.
 Nuclear imaging including radiation dose and PET scanning.
All these techniques build upon the physics content of PH2, PH4 and PH5. The study of this
option provides an ideal synthesis of many of the ideas developed especially in the A2
course.
X-rays
The nature and properties of X-rays, and their production in an X-ray tube are covered in
PH2.3 (e) and (f). The following diagram describes their production in sufficient detail.
Lead
shielding
Very high
voltage
Copper/
Tungsten
block (anode)
Heater
Curren
t
supply
Heater
filament
Electron
beam
Focussing
anode
X rays
Cooling fins
The heater ‘boils off’ electrons by thermionic emission. These are then accelerated to very
high velocities by the p.d. between the heater filament and anode. They are collimated by the
focussing anode. The tube is evacuated so the electrons travel in straight lines and collide
with a tungsten target (the anode) embedded in a copper block. The resulting deceleration
produces an enormous amount of heat (up to 99% of the energy input) and also X-rays, which
emerge from a window in the lead housing.
A continuous spectrum, then, can be obtained by electrons decelerating rapidly in the target
and transferring their energy to single photons. This radiation is known as ‘Bremsstrahlung’
or braking radiation. Superimposed on the continuous spectrum are several sharp lines. These
result from the bombarding electrons knocking out orbital electrons from the innermost shells
of the target atoms. Electrons from outer shells will then make transitions to fill the gaps in
the inner shells, emitting photons whose energies are characteristic of the target atom.
26
Transitions into the K shell give rise to K lines, the L shell L lines and so on. For heavy metal
targets the resulting photons are in the X-ray range. A typical intensity spectrum would be:
1.0
0.5
λmin
0.05
0.1
0.15
The intensity of X-rays is defined as the energy per second per unit area passing through a
surface. This can be increased by increasing the voltage of the X-ray tube, or by increasing
the current supplied to the filament. The photon energy is also determined by the tube voltage
with the maximum photon energy being given by:
hc
Emax 
 eV ,
min
where V is the tube voltage. The optimum photon energy for radiography is around 30 keV
which is obtained using a peak tube voltage of 80 – 100 kV.
A narrow beam of X-rays is preferred as this reduces scattering and so leads to a sharper
image. Blurring can also occur due to scattered radiation. This can be reduced by introducing
a grid directly in front of the detector. This grid consists of a large number of lead strips so
that only primary or direct radiation will get through to the film.
X-ray beam
Patient
Scattered
Radiation
Transmitted
Radiation
Grid
photographic
film
27
When X-rays pass through matter they are absorbed and scattered and therefore the beam is
attenuated. This attenuation can be calculated using the equation:
I  I 0e  x
Where I = intensity at a depth x
Io= intensity at the surface
µ = the attenuation coefficient
Note that the half value thickness, x½ can be calculated using
ln 2
.
x12 

This can be derived in the same way as the half life equation in radioactivity.
The main absorption mechanism of X-rays in the body is the photoelectric effect. The X-ray
photon is absorbed by an electron which then leaves the atom. This is more efficient for
atoms with larger numbers of electrons i.e. higher atomic numbers. Consequently denser
materials such as bone will absorb more X-ray photons than less dense areas such as soft
tissue. This will lead to a large contrast between bones and soft tissue and therefore a sharp
image. If there is not a great contrast between the areas of the body being studied then
sometimes a contrast media is used, e.g. a barium meal when studying the stomach or
intestines.
Computed tomography (CT or CAT scan) also uses X-rays, but in this case the X-ray tube
moves in a circle around the patient taking images of the body at all different angles. A
computer combines these images to produce a cross sectional image of the body. By adding
these slices together a 3-D image can be produced.
CT scans are very quick to produce and show a wide range of different tissue clearly. They
do however subject a patient to a high dose of radiation and the machines are very expensive.
Ultrasound
Ultrasound can be generated using piezoelectric crystals. If you apply an alternating p.d.
across the crystal you cause it to become deformed, with the crystal vibrating at the same
frequency as the applied p.d. This can be used to generate ultrasound. The process also works
in reverse, with the crystal receiving ultrasound and converting it to an alternating p.d. The
crystal, then, can act as both an emitter and a receiver of ultrasound.
Ultrasound can be used in diagnosis in two different ways
1. A-scans, or amplitude scans, where a short pulse of ultrasound is sent into the body and a
detector (usually connected to a C.R.O.) scanning for reflected pulses. Using the time
base, the time the ‘echo’ takes to return can be found and the distance between structures
in the body can calculated. A-scans are usually used when the anatomy of the section is
well known but the precise depth is needed e.g. a delay in measuring a known position in
the brain could indicate the presence of a tumour.
2. B-scans, or brightness scans in which the reflected pulse is displayed by the brightness of
the spot. If an array of transducers is used a 2-D image can be built up. This is widely
used to assess the health and growth of a prenatal foetus.
28
Acoustic impedance is defined by the equation
Z  v
Where Z = acoustic impedance of the medium
Ρ = Density of the medium
v = Speed of sound in the medium
The acoustic impedance determines how much energy (ultrasound) is reflected at a boundary.
If two materials have a large difference in Z then a lot of the energy will be reflected back,
but if there is little or no difference in Z then there will be little or no reflection.
For soft tissue such as skin the acoustic impedance is very different from that of air, so, if an
ultrasound probe were simply placed upon the exterior of the body, most of the ultrasound
would be reflected rather that entering the body. To prevent this, a coupling medium such as
a gel or oil which has a Z value much closer to that of soft tissue must be first applied to the
patient. This removes the thin layer of air and enables a much greater percentage of the
ultrasound to enter the body.
Blood flow around the body can be studied using a Doppler Ultrasound probe. The Doppler
Effect (PH4.5) is the shift in frequency of a wave when it received by a moving object (either
towards or away from the source). It also occurs when a moving source sends out a wave. A
continuous source of ultrasound is sent out and its echo received back off a moving object
such as blood cells. Because the blood cells are moving, the Doppler shift is doubled: once
for the blood cells receiving the ultrasound and once for the blood cells acting as transmitters
back to the stationary receiver. As the ultrasound is sent out continuously, two transducers are
needed: one to produce the ultrasound; the other to receive it. By analysing the shift in
wavelength of the received wave ∆λ and comparing it to the initial wavelength λ the velocity
can be calculated using the equation:
 2v

,

c
where c = the speed of the ultrasound wave. Note the 2 in the equation.
This technique will show up any changes in the blood flow through a vein or artery and so
can be used to detect clots or thrombosis.
Magnetic Resonance Imaging [MRI]
Nucleons (protons and neutrons) possess spin, which makes them behave like small magnets.
Usually these will cancel each other out. However a hydrogen nucleus only has one proton
and it is this nucleus that is studied using an MRI scan. Under normal conditions the
hydrogen nuclei will be randomly arranged and cancel each other out. However if a strong
magnetic field is applied they tend to align themselves, in almost equal numbers, either with
the field lines (in parallel) or exactly opposite to the field (antiparallel).
The nuclei are in continuous motion, due to thermal energies, and will all wobble or precess
around the field lines at the same frequency (called the Larmor frequency)
If radio waves are directed at the hydrogen nuclei at the same frequency as they are
precessing (Larmor frequency), they will resonate and flip from one alignment to another so
producing a magnetic field. When the radiowaves are switched off the nuclei revert back to
their original state giving off electromagnetic radiation. It is this signal that is detected by the
scanner. The time taken for the nuclei to switch back is called the relaxation time, and
depends on what tissue type the nuclei are in. By measuring the various properties of the MRI
signal along with the relaxation time a detailed image of a cross section of the body can be
built up.
29
Magnetic resonance imaging is particularly good in obtaining high quality images of soft
tissue such as the brain, but is not as good for harder objects such as bone.
Some of the advantages and disadvantages of X-rays, ultrasound and MRI for examining
internal structures can be summarised by the following table:
Technique
Advantages
Disadvantages
X-rays
X-rays are absorbed by bone
and so produce good shadow
images.
Unlike ultrasound they can
produce images of, e.g. cancer
on the lungs.
High radiation dose for the
patient.
People working with X-rays
need to take care to limit their
annual dosage.
Ultrasound
No known side effects.
Good quality images of soft
tissue.
Moving images can be
obtained.
Machines are relatively cheap
and portable
Doesn’t penetrate bone and
so cannot study the brain.
Cannot pass through air and
so cannot study the lungs.
Low resolution.
MRI
No Known side effects.
High quality images of soft
tissue.
Image can be made for any
part/orientation of the body.
Images of hard tissue such as
bone are poor.
Uncomfortable
for
the
patient,
causes
claustrophobia.
Very expensive.
30
Electrocardiograph (ECG)
The heart is a large muscle which acts like a double pump. The left hand side receives
oxygenated blood from the lungs and pumps it to the rest of the body. The right hand side
pumps the blood returning from the body, at low pressure, to the lungs.
The heart typically beats at between 60 and 100 times a minute. Each beat is triggered by a
pulse starting in the upper right region by a cluster of cells called the sinoatrial node. This
signal spreads through the atria causing them to contract, forcing blood into the ventricles. A
short time later the electrical pulse reaches the ventricles causing them to contract forcing
blood out of the heart. There is a one way valve between the atrium and the ventricle to
ensure the blood flows the right way.
31
A typical ECG recording consists of three parts:
1. The P wave which occurs because of the contraction of the atria;
2. The QRS wave which is due to the contraction of the ventricles;
3. The T wave which is due to the relaxation of the ventricles.
It is the signals from the sinoatrial node which are detected by the ECG. This signal is very
weak by the time it reaches the body’s surface. The electrodes ( to detect the signal) need to
be placed on the limbs and chest, where the arteries are close to the surface, and also need to
be attached with conducting gel after all body hair has been removed. The signals also need
to be heavily amplified once received.
Any deviation from a ‘normal’ ECG indicates some form of cardiac disorder. This can be
used to look for muscle damage (heart attacks), irregular pumping (arrhythmia), blockages in
the heart due to disease and the heart going into fibrillation, where the beating is fast and
irregular.
Nuclear Imaging
The effects of α, β and γ radiation on living matter will be examined in the same depth as in
PH5.4.
The unit of absorbed dose for matter is the gray (Gy), where
1Gy = 1 joule per kilogram
Because ,  and  radiation interact differently with living tissue, the gray does not
adequately describe the effects of the different radiations. 1 mJ kg1 of  particles has a much
greater biological effect that X-rays or -rays because the radiation is so heavily ionising and
therefore the range of the  particles is so low. To allow for this, a quality factor Q is used:
For many tissues a value of Q = 20 is used for  particle and Q = 1 for ,  and X-radiation.
Different tissues also respond differently and hence different Q-factors are employed by
medical physicists. With this quality factor, the quantity is referred to as the “dose
equivalent” and its unit, the sievert (Sv).
For  and X-rays the dose is equal to the dose equivalent. In general:
Dose / Sv = Q  Dose equivalent / Gy
In examination questions, the value of the Q-factor will always be given.
A gamma camera uses tracers to produce images. A common nuclide used is technetium 99. This is attached to a molecule that will be taken up by the tissue to be studied. It has a
half life of about 6 hours, which is long enough to allow the nuclide to be transported to the
site of interest and for the radiation to be studied but short enough so that the nuclide not stay
active inside the patient for too long.
Once the isotope has been absorbed by the body the gamma rays are directed towards a
crystal (usually sodium iodide with a small amount of thallium added) through a lead
collimator, which consists of a lead circle which has a regular number of holes drilled in it.
This absorbs all the gamma rays that enter the collimator at an angle. The gamma photons
that pass straight through the collimator hit the crystal and cause it to scintillate. The gamma
ray excites electrons in the crystal causing them to give off visible light. This light is detected
by a bank of photomultiplier tubes which build up an image of the levels of gamma radiation
being emitted from different parts of the tissue.
32
Positron Emission Tomography (PET) Scanning
This requires the use of an isotope that emits positrons (beta+ particles). These will only
travel very short distances before slowing down and attracting an electron from a nearby
atom. These will annihilate each other giving off two gamma rays, of identical wavelength,
which will move off in opposite directions (thus conserving momentum).
The PET scanner detects these gamma rays when they reach a scintillator in the scanning
device. This produces a small flash of light which is picked up and amplified by a series of
photomultiplier tubes in a similar way to the gamma camera. This technique depends on the
simultaneous detection of a pair of gamma photons. Individual gamma photons are ignored.
In order to produce positron emitting isotopes a cyclotron is needed (spec reference PH5.2) as
they have very short half lives. This is very expensive. PET scans are now often combined
with CT scans which enable both soft and hard tissue to be seen clearly.
REFERENCES
Advanced Physics Adams and Allday Oxford press
Advanced Science Physics P. Fuller Heinemann
Medical Physics M. Hollins University of Bath Science Nelson
Medical Physics J.A.Pope Heinemann
WEB REFERENCES
www.wikpedia.org Medical Physics
www.teachingmedicalphysics.org Institute of Physics
www.s-cool.co.uk
www.antonine-education.co.uk
www.medphys.ucl.ac.uk
33
Unit PH5 Option E – Energy Matters
The aim of this option is to allow students to explore arguably the most pressing topic of our age in
the context of underlying physical principles. After studying this option, students will be in a position
to absorb and dissect information, often contradictory and misleading which is presented in the
popular media and to make informed decisions. Energy already pervades much of the specification.
The Table below identifies these earlier energy-related topics and gives an overview in the centre box
of how they are developed, extended and linked in this option.
PH2.5
BLACK BODY RADIATION.
STEFAN & WIEN LAWS
PH1
MECHANICS
ENERGY
RENEWABLES
TIDAL
HYDROELECTRIC
WIND
STORAGE
SOLAR ENERGY
GREENHOUSE EFFECT
SAVING
(Thermal Conductivity)
STOCKS
HAZARDS
WIDER ISSUES:
(Economic, Social, Political)
WORK FROM HEAT
SECOND LAW
HEAT PUMP
FUEL CELL
PH4
THERMAL PHYSICS
FISSION
Enrichment
Breeding
PH5.5
NUCLEAR
PHYSICS
FUSION
Problems & Prospects
(JET, ITER)
PH1.5
SUPERCONDUCTIVITY
PH5.2
B FIELDS
34
Electricity Generation from available environmental kinetic energy and gravitational
potential energy – statements (a) – (c).
The enormous world-wide efforts now being made to increase and develop “renewable”
energy provision are the result of two factors: first, fossil fuels, on which we now chiefly
depend, are running out; secondly, these fuels are causing potentially catastrophic
environmental damage. Renewable in this context simply means “does not run out” within
the life times of human civilisation.
Renewable sources and storage systems of the kind listed in (a), (b) and (c) largely involve
applications of basic physics developed at AS (PH1 Mechanics and Energy). In the case of
wind power, the maximum theoretical power Pmax available is determined by the rate of
kinetic energy transfer through the turbine and candidates should be able to deduce (but not
remember) Pmax = ½Av3 and draw conclusions. The power derived in practice is much
below the theoretical maximum due to not all the KE being transferred, losses in the turbine
etc. For tidal and hydroelectric sources, and also storage systems, estimates can be made
from gravitational potential energy calculations based on simple models.
The importance of projects such as Dinorwig in storing energy at times of low consumption
should be recognised. The importance of renewals is nowhere greater than in the UK where
there are Government targets to increase renewals from the present 4% of our total energy
consumption to 10% by 2010 and 20% by 2020. Of the projects listed in the Specification,
the recently completed Three Gorges (Yangtze) is worth special mention because of its sheer
size, and also because of the social, economic and political implications. This is essentially a
large 1.4 mile dam with 370 miles of headwater flooding 620 square miles and having a
generating capacity of 18.2 GW – originally estimated to supply 10% of China’s energy
needs. More than a million people have been displaced in its construction – an exercise more
easily accomplished in China than in most countries. This is an extreme case; in all major
projects there are negative factors: noise, disruption, use of land, effect on wildlife etc and
almost always there are protest groups.
Though the physics of all the above is straightforward, the handling and conversion of the
various units can be troublesome. The kilowatt-hour conversion 1 kWh = 3.6 MJ is needed
frequently and worth remembering while familiarity with the SI prefixes M, G, T, P and E
will be found helpful. Some National and International data tables give energies in mtoes
(million tonnes of oil equivalent). The conversion is 1 mtoe = 42 TJ approximately and this
will be provided in any question.
35
Nuclear sources – nuclear equations, fission, breeding and enrichment
Uranium derived from mined ore contains only 0.7% of U235, the isotope which fissions
with thermal neutrons and is the main source of nuclear power. U238, which makes up the
remaining 99.3%, does not fission with thermal neutrons though it does make a small energy
contribution by limited fission with fast neutrons. The ideal fission reactor would be fuelled
by pure U235, but it extremely difficult to separate the isotopes as their masses are so nearly
the same. The current method is by ultracentrifuge. Vessels containing Uranium hexafluoride
gas are spun at enormous angular velocity. Because the two isotopes have different masses
they will experience (slightly) different radial forces with the result that the heavier U238
tend to concentrate on the outside of the vessel and the lighter U235 nearer the axis. The
process is slow and the apparatus complex and costly. In practice reactors running on
“enriched” Uranium, in which the U235 has been increased to just a few percent, are more
efficient and cost effective so that the costs of enrichment are more than recovered. Also,
pure Uranium235, obtained through repeated enrichment, is used in the fission bomb which is
a cause of unease when new nations embark on enrichment programs. The remaining U238
after the U235 has been separated is known as depleted uranium (DU). It is much less
radioactive than the original uranium, because the half-life of U-238 is so long [4.5  109
years] and because of its high density has been used lately in armor penetrating shells amid
some controversy [PH6 resource folder 2003].
U238 can itself be made to produce useful fissionable material by “breeding”. U238 captures
fast (high-energy) neutrons to form unstable U239 which decays to Np239 by  emission
which in turn decays to Pu239 by further  emission . (A useful exercise might be to write
down the equations involved in this sequence given that Z = 92 for Uranium). Pu239 is an
isotope of plutonium which does not occur naturally [or, more strictly, any Pu-239 originally
present in the primordial solar nebula has long since decayed]. It turns out that Pu239 fissions
with thermal neutrons similarly to U235 and can therefore be used as a primary source of
nuclear energy. Reactors which produce Pu239 in this way are known as Fast Breeder
Reactors – fast because fast or high energy neutrons are required to start the process. These
reactors have the materials and moderators so arranged that enough fast neutrons are
absorbed for breeding while sufficient moderation occurs to provide thermal neutrons to
sustain the U235 chain reaction. The eventual separation of plutonium from uranium is
relatively easy (compared with U235 and U238 separation) because they are chemically
different. Note that fission produces radioactive waste which must be safely disposed of (still
a major problem), that there are risks (Chernobyl), that stocks of ore are limited (therefore
non-renewable) but that there are no carbon emissions other than that involved in the initial
building of the necessary infrastructure.
Nuclear fusion
It is important to know the DT (deuterium-tritium) reaction and why it is the most suitable for
terrestrial fusion (timescale and temperature). It is useful to look again at the proton chain in
PH2.5 and to be reminded why this would not work on earth. An excellent account of nuclear
fusion is given on the Joint European Torus (JET) website www.jet.efda.org . The key points
to note are:
 the kinetic energies of the particles must be greater than the Coulomb interaction for
interaction to occur meaning temperatures of 108 K;
 plasma containment at this temperature only possible by a combination of magnetic
fields – the tokamac;
36


very strong B fields are required to deflect and therefore contain fast particles (see
PH5.2(m) for background);
fields of around 5 T require currents of around 7 MA which are only really achievable
with superconducting coils (see PH1.5 (k) to (p)).
The problems are immense and continuous energy through fusion is still a long way off. But
huge efforts and investments are being made. The International Thermonuclear Experimental
Reactor, ITER ( www.iter.org ), a joint venture by most of the great Nations –including
China, India, USA, Russia, EEC – is underway in France and due to power up in 2016.
Fusion, if it works, will provide all our energy needs. Deuterium is abundant in seawater [~ 1
in 104 hydrogen atom is deuterium] and tritium can be obtained by neutron capture by lithium
in the reactor itself. There are no toxic products.
Heat transfer processes – convection and conduction
These are familiar topics well covered in standard textbooks. Emphasis will be on thermal
insulation and energy-saving, but note that publications on these topics use Thermal
Transmittance, (U value) rather than thermal conductivity K. The U value of a slab of
thickness d is given by
U = K/d.
This will be provided, if required, in any question.
Solar radiation as an energy source
A form of renewable energy quite different from those treated earlier is solar energy. The
background physics has already been developed in PH2.5(a to d). Revision of this material is
a good starting point with emphasis on the laws of Wien and Stefan, the inverse square law
and what is meant by black body.
A key quantity is the Solar Constant – the total radiated power per square metre crossing a
plane perpendicular to the earth-sun radius measured just outside the earth’s atmosphere. The
value is not constant (despite the name), as the earth-sun distance varies over the year, but
averages at 1.35 kW m-2. We can estimate the rate of solar power arriving at earth, ignoring
clouds, atmospheric absorption etc. as of the order of 1017 W and compare this with the rate at
which energy is consumed throughout the world (of the order of 1013 W). So there is
abundant solar energy; the problem is harnessing it effectively. There are two ways:
 solar panels;
 photovoltaic cells.
In the solar panel, water is heated directly from sunlight. The panel contains a flat coil of pipe
connected to the domestic hot water cylinder and is placed, ideally, on a South-facing roof.
As the name suggests, the photovoltaic cell produces electrical energy from solar energy. At
present, photovoltaic cells make very little contribution to our energy because of high
manufacturing cost. Most of the cells currently in use are made of very pure silicon which has
to be doped and cut in a special way – all very expensive.
Not for examination: As semiconductor devices and band theory are not in the specification, this
rough outline may be helpful: the silicon cell consists of n- and p-doped regions forming a p-n
junction. Incident photons excite electrons into the conduction band creating electron-hole pairs
which migrate to form an electric current. Detailed knowledge will not be expected.
37
Much work is in hand to develop cheaper and more efficient cells using, for example,
composite materials. The Energy Conversion Efficiency of a photovoltaic cell is defined by
the usual efficiency equation, in this case written as:
Conversion Efficiency 
Useful energy extracted
100%
Total energy input
Values range from around 6% for the cheapest commercial cells to around 40% for the most
expensive state-of-art cells. It follows that large areas of cell are required for even moderate
power such as typical domestic consumption, but small cells of around 5 cm2 are sufficient to
power pocket calculators which require less than 1 mW.
Carbon footprint
Candidates can not be expected to remember statistics, but some key figures are worth
bearing in mind. For example, about one fifth of UK electricity is from nuclear reactors and
three quarters from fossil fuels (coal, oil and natural gas). Fossil fuels have two major
drawbacks. First, the stocks are finite. Secondly they produce carbon dioxide gas which is
harmful if allowed to escape into the atmosphere. In a chemical reaction in which carbon is
oxidised (burned), each carbon atom combines with two oxygen atoms from the atmosphere
to form carbon dioxide CO2. Straightforward calculation from the atomic masses shows that
one kilogram of carbon produces 2.66 kg of carbon dioxide.
Carbon dioxide is a “greenhouse” gas and its increased presence in the atmosphere leads to
global warming in the following way:
 The solar radiation spectrum covers a range of wavelengths with maximum power at
around 480nm which is at the blue end of the visible spectrum. This value is
determined by the temperature of the sun’s surface (Wien’s displacement law;
λmax T-1).
 The atmosphere is essentially transparent to this wavelength so the solar energy
passes through and is absorbed at the earth’s surface.
 The earth in turn radiates thermal energy but, because the earth’s surface temperature
is much lower than that of the sun, this peaks at around 10 m which is in the far
infrared region.
 Carbon dioxide absorbs strongly at this wavelength, and re-emits in all directions
including back to earth leading to global warming. Other polyatomic molecules such
as methane and nitrous oxide behave similarly but carbon dioxide is more abundant.
This is how greenhouses heat up – hence the name; glass, like CO2 is transparent in
the visible but absorbs in the IR.
Experiments show that burning 1 kg of carbon produces about 13 kWh of energy which
works out at around 6 eV per carbon atom. It is interesting to compare this with the 200 MeV
produced by the fission of one U235 nucleus.
The consequences of increasing greenhouse gases need to be recognised: global warming,
polar icecap melting, weather changes, flooding, more hurricanes etc. Also important is the
decline of vegetation, particularly the rain forests, which remove CO2 from the atmosphere
through organic growth. Important too is recognition that the worlds population is increasing
as is the industrialisation (and hence energy requirements) of emerging nations: also to be
noted are the measures to counter the ill effects (national and international reduction targets,
Kyoto protocol, “carbon footprints”, Environmental Impact ratings etc).
38
Fuel cells
Interest is reviving in fuel cells as they offer the possibility of efficient and environmentally
friendly energy production, especially for transport, and could become a replacement for the
internal combustion engine when the oil runs out. Prototype cars powered by fuel cells
already exist. The fuel cell is electrolysis in reverse. As electrolysis has long disappeared
from physics specification, a brief outline is necessary.
When an electric current is passed through water, ionization of the water molecules occurs
through collision with the charge carriers. Avoiding the detailed chemistry, the upshot of this
is that water is broken down into its constituent gases with hydrogen bubbles collecting at the
cathode and oxygen at the anode. The apparatus is simple: a dish of water with two electrodes
each with an inverted jar above it to collect the gases, and a current source. A drop of acid is
needed to make the water conducting and the anode made of platinum to avoid oxidation.
So, in summary, electrical energy breaks water down into oxygen and hydrogen gases. Can
the reverse take place in which hydrogen is recombined with oxygen to provide electrical
energy? Fuel cells do just this. The process is complex but in crude outline the following is
what happens in the Polymer Electrolyte Fuel Cell (PEFC):
 Hydrogen is supplied to the anode where the atoms are ionized by a catalyst.
 A polymer electrolyte then routes the electrons to the cathode via an external circuit
forming a useable electric current.
 The protons continue through the polymer electrolyte to the anode where they
recombine with the electrons, and the hydrogen reacts with oxygen, which is fed
directly into the anode, to form water.
One great advantage is that there are no damaging products – particularly no carbon dioxide.
Another great advantage is that there is no heating – useful energy is not being obtained from
heat; this will be returned to later after Heat Engines. Also the cell can be connected directly
to electric motors on drive wheels of cars, cutting out the heavy and inefficient engines
(cylinders, reciprocating pistons etc) of traditional cars. Hydrogen of course is a hazard and
there are difficulties over its delivery and storage, but its supply will become plentiful if
fusion succeeds – hydrogen and oxygen produced by electrolysis of water from electricity
from turbines powered by fusion reactors – but this is a long way off. At present, the
experimental fuel cell cars are extremely expensive and the electrolyte polymer degrades and
has to be replaced within the lifetime of the car.
A useful website, albeit commercial is:
http://automobiles.honda.com/fcx-clarity/how-fcx-works/v-flow/
39
Thermodynamics – Carnot cycles, heat pumps and the 2nd Law of Thermodynamics
The groundwork for understanding the central problem of obtaining useful work from heat
has been developed already in PH4.3 Thermal Physics (i) to (d) and a good starting point is to
look again at these. One ideal heat engine consists of a cylinder with a piston and containing
ideal gas (the working substance). The engine is also “ideal” in the sense that the walls and
the piston are perfect heat insulators, the base is a perfect conductor and there are no friction
losses as the piston moves. The process involved is treated in many textbooks but is set out
here for convenience.
1. The engine is first placed on heat reservoir at temperature T1; initially the state of the
gas is shown by point A on the p – V diagram.
Heat Q1 passes from the reservoir into the gas as the state of the gas changes from A
to B. Work is done as the gas expands (piston goes up) but the internal energy does
not change as the curve AB is an isotherm (temperature constant so internal energy
constant). The changes along are given by the First Law of Thermodynamics (in the
form U = Q  W ) and are shown in the first line of the table.
2. The engine is now transferred to a perfectly insulating stand and allowed to expand
further to state C; no heat is transferred but work is done (no need to mention
adiabatic); changes are given in line 2.
3. The engine is now transferred to second reservoir at T2 and the gas is compressed
isothermally to state D; changes are given on line 3 but note particularly that heat is
ejected from the gas into the second reservoir – the sink.
4. Finally the cylinder is placed back on the insulating stand and compressed to return to
state A. The cycle is complete, and we can see from the bookkeeping table that an
amount of work (Q1-Q2) has been obtained from an amount of heat Q1.
Heat
reservoir
T1
T2
BC
AB
CD
Step
Q
U
W
A to B
Q1
0
Q1
B to C
0
(U1U2)
U1U2
C to D
-Q2
0
Q2
D to A
0
U1U2
(U1U2)
Heat
sink
DA
40
pressure, p
A
Indicator diagram
for a Carnot cycle
Q1
D
B
T1
Q2
C
T2
Volume, V
The procedure can be summarized as shown below:
HOT
T1
HEAT Q1
ENGINE
WORK
W = Q1  Q2
HEAT Q2
COLD
T2
The question arises, “Why does heat have to be ejected?” Or, equivalently, “Why does there
have to be a sink?”
If the process terminated at B the engine would be of no further use. To obtain useful work
the process must be continuous, which requires that the cycle be repeated over and over
again for as long as is necessary. This means repeatedly returning the gas to its original state
– something which can only be achieved by ejecting heat at one stage in each cycle.
The efficiency of the heat energy is defined by:
Useful work output
Efficiency 
100% .
Total energy input
Q  Q2
Q
Efficiency  1
 1 2
So
Q1
Q1
(1)
41
All that we know, or need to know, about the source and sink is that they have temperatures
of, respectively, T1 and T2. Clearly the amounts of heat transferred will be governed by these
temperatures. In fact, the ratio Q2/Q1 is used to define the ratio of Kelvin temperatures so
that
Q2 T2

Q1 T1
and equation (1) above becomes
T
Efficiency  1  2
(2)
T1
So a Carnot engine working between 100oC and 0oC [373 K and 273 K] can not have an
efficiency of greater than 27%. In practice there will be other factors such as friction and heat
loss which will make the actual efficiency even less. In the ideal engine described here there
are no such losses, and the cycle is therefore reversible in that if the operations described are
performed in the opposite (anticlockwise) sense the system is returned to precisely its initial
state. This would not happen if there were losses.
All this shows a fundamental difference between heat and other forms of energy. While
electrical energy can be transformed entirely into heat (I2R heating in a resistor) only a
fraction of heat can be transformed into useful work. This has been shown here to be true for
the ideal Carnot engine; it is in fact generally true and is formulated in the Second Law of
Thermodynamics which will be looked at later.
Details of actual engines, to which the same principles apply, will not be expected. The
simplest example is probably the steam engine, but this is now of historical rather than
practical interest. The Otto cycle for the internal combustion engine is worth looking at as a
more complex example.
The running of the Carnot cycle in reverse has already been mentioned and this is the basis of
the refrigerator and the heat pump. They operate on the same principle – that of extracting
heat from a cold source and ejecting it at a hot sink – and the Carnot cycle runs anticlockwise around the indicator diagram. Work is required to achieve this and a schematic
diagram for the process is given below. Note that the same diagram applies both to the
refrigerator and the heat pump, but that we do not use the term “efficiency” to describe their
effectiveness – efficiency is reserved for the heat engine. Instead, the figure of merit for these
devices is the Coefficient of Performance. The definition of the CoP is essentially the same as
that for efficiency:
Useful energy transfer
100% .
Work input
The nature of the “Useful energy transfer” differs however and the CoP is defined separately
for each case below.
i.e.
Coefficient of Performance 
42
For the fridge:
HOT
T1
Heat extracted from cold source
100%
Work input
Q
Q2
 2 
W Q1  Q2
COP 
HEAT Q1
ENGINE
HEAT Q2
COLD
T2
WORK
W = Q1  Q2

T2
T1  T2
For the heat pump:
Heat delivered to hot sink
COP 
100%
Work input
Q
Q1
 1
W Q1  Q2

T1
T1  T2
For refrigeration, the source is the inside of the fridge and the sink is the kitchen or, more accurately,
the cooling grill at the back of the cabinet. In the case of the heat pump the source is the ground,
preferably in ground water or a river for better heat transfer, and the sink is the radiators inside of the
house to be heated. Practical details are not expected but the broad principles should be understood.
For example, the working substance or “refrigerant” might be a liquid which can be made to
evaporate at the cold source, thus absorbing heat, and then to eject heat at the hot sink where the vapor
condenses back to liquid. Work is done in circulating the refrigerant and bringing about the necessary
phase changes.
The COPs are, like the Carnot efficiency, theoretical maximums. In practice, performance is much
poorer owing to the usual losses, but it is still instructive to insert numbers for typical situations. Heat
pumps seem a good proposition, but the capital expenditure is high and the losses great. They are used
in major buildings (Festival Hall, Buckingham Palace, and more recently the Senedd at Cardiff) while
many firms offer domestic appliances (try inserting “Heat Pump” into a search engine).
As already mentioned, the limitations of heat into work are a consequence of the Second Law of
Thermodynamics, the Kelvin statement of which is: “No process is possible the only result of which is
the total conversion of heat into work”. To illustrate this, look again at the Carnot cycle and the step
from A to B. This would seem at first to contradict the law for all the heat absorbed in the step is
converted into work. But this is not the only result because, in the process, the state of the gas
(pressure and volume) has changed. The law is fundamental, far reaching and one of the great
cornerstones of science. For example, an essential stage in obtaining work from a nuclear or any other
power plant is that of transforming heat to drive a turbine, and all the limitations of the Carnot cycle
apply. Suppose steam enters the turbine at 500 oC and then condenses to water at 60 oC; the Carnot
efficiency is 0.57 and the actual efficiency will be much less again. So, most of the energy is wasted.
One way to recover some of this is to pump the hot water from the power station condenser around
local housing in a massive central heating system. This is known as a Combined Heat and Power
(CHP) scheme and was pioneered in Britain at the old Battersea power station.
43
The overall efficiency obtained in a CHP scheme is much greater than the Carnot efficiency because a
large fraction of the “waste” heat Q2 now becomes useful – though there will be some losses between
the power station and the houses. In practice, CHP generators are run with a higher temperature cold
sink [~ 100C], thus reducing the efficiency of the electricity generation but allowing for more useful
heat energy distribution. It is easy to understand now why electricity should not be used for heating –
most of the original energy has already been wasted as heat. We can also now better understand a
great advantage of fuel cells – no heat is involved so there is no Carnot wastage in the cell itself,
though heat may have been involved at an earlier stage in producing the hydrogen and oxygen.
GCE Physics – Teacher Guidance
4 December 2007
Download