Capacitors, Inductors, RC and RL Circuits

advertisement
Module 2: Capacitors, Inductors, RC and RL Circuits
I.
Introduction: Time Dependence and Circuits
So far, everything we have considered has been “static” in that we have ignored time dependence. We
solve for V’s, I’s and P’s assuming they are constant in time. Further, when circuits are changed (i.e.
“opened” or “closed” using a switch, or a component is replaced) we assume that the currents, voltages
and powers transform instantaneously from their “pre” values to their “post” values (i.e. when you close
a switch the current goes immediately from zero to its final value).
Well, we know that real life is much more complicated than that. Most interesting phenomena – and
circuits are no exception – have non-trivial time dependence. And when changes occur, they are never
instantaneous, and in fact take a finite time to occur. For example, in physics class a ball bouncing off a
wall is modeled as a projectile whose velocity goes from +v to –v instantaneously (and discontinuously).
But if v changes discontinuously, that implies that acceleration (a = dv/dt, and the applied force F = ma)
must be infinite! So a real physical model takes into account the fact that the velocity must gradually
transform from +v to –v, though in most cases this happens quickly enough that the assumption of
instantaneous velocity change gives effectively the same answer as you get with a more complex model.
However, this is by no means always the case, and it’s important to recognize when this is true.
In circuits, these non-instantaneous changes are due to the fact that circuits not only are resistive to the
instantaneous flow of current, but are also reactive to changes in current or voltage. In circuits, these
“reactive” devices are capacitors and inductors. (And remember the word “react,” as it’s going to come
back when we talk about AC circuits!)
II.
The Capacitor
We start from a simple model of a parallel-plate capacitor, which is
composed of two plates of area A, separated by distance d. On one
plate a charge +Q is stored, with an equal but opposite charge –Q on the
other. (Note that when we say “a capacitor has charge Q,” what we
actually mean is that the total charge is zero: +Q on one plate, and –Q
on the other!) Since each plate has area A, the charge density on each
plate is given by
σ = ±Q/A
where σ here is charge density (not to be confused with the conductivity σ introduced in Module 1A!).
We can use Gauss’ Law ∫∫E●dA = Qenc/∈0 to show that the electric field in the region between the
plates is uniform (provided A >> d, so that fringing effects are minimal) and given by
E = σ/∈0
where ∈0 = 8.85 x 10-12 C2m2/N is the permittivity of free space. It follows that the potential difference
between the plates is given by
V = ∫E●ds = E d = σ d / ∈0 = Qd/(∈0A)
Now, we may be tempted to write an “Ohm-like law” that removes all this “microscopic” stuff about
electric fields and charge densities (just as we removed all that “microscopic stuff” about collisions and
electron masses when we discussed resistivity) to give an expression which describes a relationship
between our circuit-analysis mainstays, V and I. Unfortunately, we cannot get an expression as simple
as V = I R, since the relevant quantities in the above expression are Q and V. So, we get
Q = (∈0A/d) V
→
Q=CV
Where C = ∈0A/d is defined as the capacitance. While you can use the above equation to determine
that the unit of capacitance is C/V, capacitance (like resistance and voltage) comes up frequently so it
has its own unit, the Farad (F), where 1 F = 1 C/V. However, the Farad is in practice an enormous unit,
so for the most part off-the-shelf capacitors are sized in µF or even pF. (While Farad-sized capacitors do
indeed exist, they’re either enormous – where’s your forklift? – or involve significant design trade-offs
which compromise other properties in favor of increased capacitance, so they have very limited uses.
At this point it’s worth taking the time derivative of Q = C V, to yield
I = C dV/dt
which is an equally good equation representing the behavior of a capacitor, so that we now have
expressions for both of our mainstays, V and I (and their friends Q and dV/dt), although there is not a
simple linear relationship as there was with the resistor. Given that V (and its time derivative), and I
(and its time integral) appear in the capacitor’s defining equations, it is reasonable to expect that V and I
may exhibit time dependence in a capacitive circuit.
Before moving on, let’s consider the energy stored in a capacitor. When a capacitor is charged, there is
an electric field between its plates which stores energy that can be harnessed later, when the capacitor
discharges. The energy dU required to add a charge dQ to a capacitor at a voltage V is dU = VdQ, where
V = Q/C. Of course, as you add more charge, V increases, so to calculate the total energy U, you must
integrate from the first charge added (i.e. Q = 0) up to the final charge Q:
U = ∫dU = ∫V(Q) dQ = ∫(Q/C) dQ = Q2/2C
We can also rewrite this, using the relationship Q = CV, to get
U = (1/2) CV2
As a capacitor charges (i.e. Q and V increase) an increasing amount of energy is stored in the electric
field between the capacitor’s plates. This energy is stored, and may be released later as the capacitor is
discharged.
This is an important concept, to which we will return later: while a resistor irreversibly converts
electrical energy into heat as current flows through it, a capacitor stores energy in its electric field,
which may be returned later to the circuit!
III.
The Simple RC Circuit
At right is a simple RC circuit, which consists of a resistor R
and a capacitor C connected in series. Since time
dependence is important, a switch is included, which can
be used to “open” and “close” the circuit on demand.
Let us consider that for t < 0, the circuit has been
disconnected for a long time, and that there is no charge
on the capacitor plates. Since the circuit is open, we know that for all t < 0, I = 0 in the circuit. Since the
capacitor is uncharged (Q = 0), it follows that the voltage between the capacitor plates VC = Q/C = 0.
A.
“Charging” RC Circuit
At t = 0, someone closes the switch. We can use KVL to write down an equation for the sum of the
voltage drops (let’s assume a clockwise-flowing current I, and traverse the loop in the clockwise
direction:
ε - VR – VC = 0
where VR = I R from Ohm’s Law, and VC = Q/C from the fundamental relation for capacitance. Therefore,
ε - I R – Q/C = 0
For reasons of curiosity, let’s take a time derivative of that equation. ε is constant in time, so its
derivative is zero. The derivative of I will be denoted as I’ = dI/dt, the derivative of Q is I, and R and C are
simple constant coefficients. I’ll also multiply through by a common minus sign to get:
I’R + I/C = 0
--or--
R(dI/dt) + I/C = 0
This is an equation that involves both I and its time derivative I’. You’ve probably never seen such an
equation (known as a differential equation) before, although you will soon enough! In your differential
equations class, you’ll spend a semester learning how to solve equations just like this. For now, I will
just show you a few simple cases.
B.
t = 0 behavior of Charging RC circuit
First, let’s consider what happens at exactly t = 0, right after the switch is closed. Since Q = 0, VC = 0, and
our KVL equation reduces to
ε-IR=0
→ I = ε/R
which is our familiar result from before. If we think about this a bit more carefully, this is the same
answer we’d get if the capacitor were replaced by a piece of wire (i.e. a short):
An uncharged capacitor in a circuit acts as if it were replaced by a short circuit.
Therefore, any problem (no matter how ugly!) involving only resistors and uncharged capacitors can be
reduced to one containing only resistors. And we know how to solve those, right?
We also find, by using the derivative of the fundamental capacitor equation, that
I = C dVC/dt
→
dVc/dt = I/C = ε /RC
which tells us that the voltage between the capacitor plates begins increasing with time. It also follows
that the charge on the capacitor’s plates increases with time, since C = QVC, and VC is increasing with
time.
But as VC increases, it follows that VR must decrease so that ε - VR – VC = 0, and this means that the
current I in the circuit must decrease, since Ohm’s Law tells us that the I through the resistor (which is
the only current in the circuit) is given by I = VR/R.
So the general picture is that I starts out at the value you get by short-circuiting the capacitor: I(t=0) =
ε/R. But as t increases, Q on the capacitor, and thus VC, begins increasing. As a result VR, and thus I,
must decrease.
C.
t → ∞ behavior of Charging RC circuit
If we allow this to continue for a “very long time” (which I’ll define more carefully in a moment!), the
circuit reaches a “steady-state” condition in which the V’s and I’s are constant in time. Well, if the V’s
(including VC) are constant in time, I = C (dVC/dt) implies that I = 0 (the current eventually stops flowing).
If I = 0, then VR = IR = 0, and thus VC = ε - VR = ε, and the capacitor’s charge Q = CVC = Cε. In this sense,
the “fully charged” capacitor acts like an ideal voltage source with “just the right amount” of voltage to
reduce I to zero. Equivalently,
A fully-charged capacitor (i.e. one present in a circuit in a steady-state equilibrium) acts as if it
were replaced by an open circuit.
No current flows through the capacitor, and the voltage between the terminals is equal to the value you
get from KVL by setting the current through the capacitor branch to zero.
D. What about intermediate times? Solution of the circuit differential equation
So at this point we’ve solved for the two “extreme cases:” at t = 0 the capacitor acts like a short, and at t
→ ∞ the capacitor acts like an open circuit. What about intermediate times? Here, unfortunately, we
have to get into some mathematics. Let’s return to our “circuit differential equation” obtained by
taking the derivative of our KVL loop equation:
R(dI/dt) + I/C = 0
I will rewrite this equation by doing something your calc prof probably told you never to do: I’m going
to “separate” the dI and dt in dI/dt:
dI/I = -(1/RC) dt
It may now occur to me to take a time integral of both sides:
∫dI/I = ln[I(t)] + C = (-1/RC) ∫dt = -t/RC + C’
Since there are two additive constants C’ and C that came from doing the integrals, I can replace them
both with a single additive constant C0 = C-C’:
ln[I(t)] + C0 = -t/RC
To solve for I(t), we raise e (the base of natural logarithms) to the power of both sides, and use the
familiar properties of exponents and logs:
exp (ln(I(t))+C0) = I(t)/I0 = exp(-t/RC)
→
I(t) = I0 exp(-t/τ)
where I0 = 1/exp(C0) and τ = RC represents the “time constant” of the circuit. I0 can be determined by
plugging in t = 0, in which case we see that I0 = I(t=0), which equals ε/R in our circuit. We also see that
I(t) starts at I0 = ε/R, and declines exponentially toward zero as t increases, which is exactly what we
expected from our simplistic circuit analysis.
We can now go back and solve for the rest of our circuit parameters using Ohm’s Law, the fundamental
relation for capacitance, and our KVL equation:
I(t) = (ε/R) exp(-t/τ)
(at t = 0, I(t) = ε/R, and declines toward zero with increasing t)
VR(t) = I(t) R = ε exp(-t/τ)
(at t = 0, VR(t) = ε, and declines toward zero with increasing t)
VC(t) = ε - VR(t) = ε(1 - exp(-t/τ)) (at t = 0, VC(t) = 0, but increases toward ε with increasing t)
Q(t) = C VC(t) = Cε (1 - exp(-t/τ)) (at t = 0, Q = 0, but increases toward Cε with increasing t)
An extremely useful property is that all of the behavior of this circuit can be described just using a single
number τ = RC, the “time constant.” All of the time dependence in the circuit “scales” with τ, which is
just the product of R and C. Thus we see that the “instantaneous” circuit behavior we previously
considered corresponds to the case that C = 0, so that τ = 0. But if C has a finite value (and more so if
either C or R is large), then the time dependence can be very significant!
E.
Behavior at intermediate times: Alternate derivation
Suppose we didn’t want to use the “brute force” method above of separating dI and dt? (In your
differential equations course, this is probably the first method you’ll learn, and is known as separation of
variables, because you separate your variables onto opposite sides of the equation – the I’s on one side,
the t’s on the other.)
Well, there is another, more widely-applicable method for solving differential equations. It’s known as
“guessing the answer,” or if you want to sound fancy, an “ansatz solution.” Let’s look at our original
differential equation from KVL:
R(dI/dt) + I/C = 0
If you look closely, you’ll notice that this equation has two terms – one involving a function I, and
another involving its derivative, dI/dt. Both I and dI/dt are, in principle, some horribly complex functions
of time, so both I and dI/dt can take on all sorts of values depending on what value of t you plug in.
Both terms are multiplied by only constant coefficients, yet somehow, miraculously, for any value of t
you plug in, the sum of those terms equals zero.
In other words, no matter what crazy time dependence I(t) may have, that crazy dependence is precisely
mirrored by what dI/dt does, so that the sum of the two terms is always exactly zero. This means that
both I and dI/dt must have the same functional dependence.
Well, we may ask, what types of functions return (essentially) themselves as their own derivatives?
Polynomials don’t do that. Logarithms don’t do that. Sines and cosines almost do, although you have to
take two derivatives of a sine to get back another sine (we’ll use this later when we talk about AC
circuits!). However, the derivative of an exponential is just another exponential; if we let f(t) = A exp(kt)
df/dt = d/dt ( A exp(kt) ) = A k exp(kt) = k f(t)
Let’s try plugging in this f(t) to our differential equation, using the fact that df/dt = kf:
R k f(t) + f(t) / C = 0
→
Rk + 1/C = 0
→
k = -1/RC
where the first step involved dividing by f(t) (which is okay as long as f(t) does not equal zero, also
known as the trivial solution), and the value of k represents the time constant τ we found above.
This leaves only the arbitrary constant A in f(t) = I(t) = A exp(-t/τ), which we can find using the “initial
condition” that I(t=0) = A = ε/R.
And we see that we get the exact same solutions as above.
F.
More about the time constant
One more thing we can do is inquire about what values I(t) and the other
variables take as a function of time. Since everything scales with τ, all we
need to do is compute values for one prototypical circuit for various values
of t/τ, and then “scale” the behavior for circuits with differing values of τ. I
present at right a table of values of exp(-t/τ) and 1-exp(-t/τ) for different
values of t/τ) (you can compute intermediate values using your calculator).
From this table we see, for example, that the current in a circuit drops to
0.37 of its initial value ε/R after one time constant, to 0.05ε/R after 3 time constants, and to less than
0.01ε/R after 5 time constants, for example. This gives rise to:
Rule of thumb: values may be considered to have reached their “steady state,” i.e. t → ∞ values,
after 5 time constants.
Of course, like all rules of thumb, this should be used with some discretion, since there may be cases
where that last 1% is significant! But the “5 time constant rule” is broadly applicable, and sufficient for
most purposes.
Circuit quantities that scale as 1-exp(-t/τ), such as VC(t) and Q(t), can be solved similarly. The charge on
the capacitor reaches 0.63 of its final value after 1 time constant, and the voltage across the capacitor
reaches 0.98 of its final value after 4 time constants
G. The Discharging RC circuit.
Let’s imagine that our charging RC circuit has been charging for a long time, so that all values in the
circuit reach their t → ∞ values:
I = 0, VR = 0, VC = ε, Q = Cε.
Now, the switch is opened. What happens to the charge on the capacitor?
Well, the charge has nowhere to go. The plates are, if you will, “floating” so there is no way for the
charge +Q to make its way to the opposite plate. So the capacitor will, if we ignore the effects of
leakage, remain charged with the same charge (and voltage between its plates), no matter how long we
leave the circuit there.
But I can go even further. I can remove the capacitor from the circuit, and it will continue to hold its
charge, as long as I take care not to short out the terminals. I can put the capacitor in a box somewhere,
come back later, and it will still have the same charge on its plates as it did when I put it there. (Again,
I’m neglecting the effects of leakage, which will be substantial for real capacitors.)
Or, I can take that charged capacitor and connect it up to yet another
resistor R’ (right).
As long as the switch remains open, there is nowhere for the charge on
the plates to go, so the capacitor remains charged with VC = ε (the voltage
it “inherited” from the previous circuit).
What happens when I close the switch? (For simplicity, let’s call this t = 0 yet again).
Well, once the circuit is closed, we can use KVL:
VC – VR = 0
With VC = Q/C and VR = IR’. Since at t = 0 the charge on the capacitor’s plates is not going to change
discontinuously (which would require an infinite value of I!), we can treat VC as a constant to get I =
VC/R’ = ε/R’.
As t increases and the current continues to flow, however, the capacitor begins discharging, and VC
decreases. Note that in this case dQ/dt = -I (where the minus sign represents the fact that a positive
current corresponds to a decreasing value of Q). Thus our KVL equation can be rewritten as
Q/C + (dQ/dt) R’ = 0
and we can take the time derivative of the whole equation to get (after multiplying through by a minus
sign):
I/C + (dI/dt) R’ = 0
Note that this is exactly the same equation that I solved in sections D and E, which has the solution
I(t) = I0 exp(-t/τ)
(i.e. has value I0 = ε/R’ at t=0 and declines exponentially toward zero)
Where τ is the time constant of the “new” circuit, τ = R’C.
We can now go back through, using Ohm’s Law, the fundamental relation for capacitors, and KVL to get:
VR’(t) = I(t) R’ = ε exp(-t/τ)
(has value ε at t=0 and declines exponentially toward zero)
VC(t) = VR(t) = ε exp(-t/τ)
(has value ε at t=0 and declines exponentially toward zero)
Q(t) = C VC(t) = C ε exp(-t/τ)
(has value Cε at t=0 and declines exponentially toward zero)
And this is the behavior of the discharging RC circuit.
IV.
Dealing with combinations of capacitors
So far our analysis has assumed that our RC circuit contains exactly one resistor and exactly one
capacitor in series with each other. Well, what if a circuit is more complicated than that?
If there are multiple resistors in series or parallel, we know how to deal with those -- series resistances
add, and parallel resistances combine using the “one over formula:”
Req = R1 + R2 + …
(series)
1/Req = 1/R1 + 1/R2 + …
(parallel)
Are there similar rules for combinations of capacitors? The answer is Yes.
Let’s consider parallel capacitors first. If three capacitors are connected in parallel, it follows from KVL
that all three must have the same voltage across them. Therefore, they will each carry a charge Qi which
can be calculated from
V = Q1/C1 = Q2/C2 = Q3/C3
The total charge Q carried by the three capacitors is simply the sum of the three individual charges:
Qtot = Q1 + Q2 + Q3
We can replace this network of three capacitors with one “equivalent” capacitance Ceq, with voltage V
across its plates and charges ±Qtot on its plates, with
Qtot = Q1 + Q2 + Q3 = Ceq V
→ Ceq = Qtot / V = Q1/V + Q2/V = Q3/V = C1 + C2 + C3
This can be generalized to any number of capacitors:
In parallel, capacitances add: Ceq = C1 + C2 + C3 + …
What about series combinations of capacitors? If we have three capacitors in series (depicted
at right), we know that the voltages across the three capacitors should add up:
Vtot = V1 + V2 + V3
What about the charges on the plates? There are two ways to think about this:
1.
The currents through all three capacitors must be the same at all times. If the currents are
the same at all times, then the total charge accumulated on the plates of all three must be
the same, since Q = ∫I dt.
2.
Notice that there are two “I-shaped” pieces of metal which effectively join the middle
capacitor to the two outer ones. These “I-shaped” pieces of metal are isolated from the
rest of the circuit. So if the lower plate of the upper capacitor has a charge of –Q1, and the
upper plate of the middle capacitor has a charge of +Q2, if the initial charge on that “Ishaped” piece of metal was zero, it must remain zero at later times, since there is no way for
charge to get on or off that isolated piece of conductor. Therefore, Q2 = Q1, and all
capacitors in series have the same charge on their plates.
I like to call the method in 2 “Carlo’s I-theorem,” because of the I-shaped pieces of metal. If the picture
is rotated 90 degrees, one can use its corollary, “Carlo’s H-theorem” (using the isolated H-shaped pieces
of metal that result) in the same way. As usual, the people responsible for naming things haven’t
responded to my application…
In any event, we agree that the charges stored on series capacitors are the same, and the total voltage
across the series combination is the sum of the individual voltages:
Vtot = V1 + V2 + V3 = Q/C1 + Q/C2 + Q/C3
We can define an equivalent capacitance Ceq = Q/Vtot, to get:
Vtot = Q/Ceq = Q/C1 + Q/C2 + Q/C3
→
1/Ceq = 1/C1 + 1/C2 + 1/C3
So, in parallel, you use the “one over formula” to compute Ceq.
Note that the formulas for combining capacitors are exactly the same as those for combining resistors,
except that they are reversed: you use the one over formula for capacitors in series, and the additive
formula for capacitors in parallel, rather than the obverse used for combining resistors.
Now, given some complex RC circuit, we can attempt to reduce it to the “simple RC circuit” we know
how to solve by (1) reducing our resistors down to a single Req using series and parallel combinations,
and (2) separately reducing our capacitors down to a single Ceq using series and parallel combinations.
In these cases, the circuit behavior is still the same as presented above in our “charging” and
“discharging” examples, but with τ = ReqCeq.
If, however, the circuit cannot be solved by this “decomposition” method we have to revert to our lesspreferred methods of circuit analysis.
In some cases (typically, when there is one capacitor connected to some horrid combination of power
sources and resistors) we can apply one of our linear network theorems (superposition, Thevenin or
Norton) to replace a horrible RC circuit with something less horrible, which can be solved in analogy to
our “simple RC circuit.”
In other cases, particularly where there are multiple capacitors connected in non-trivial ways, we have
to use our Kirchhoff-based methods (branch currents, mesh currents, or node voltages). In such cases,
the exponential solutions we derived above may not even be applicable, since different capacitors can
charge at different rates, and the behavior can depart significantly from the ideal exponential behavior
we found in the simple RC circuit.
However, if we do this, rather than having simultaneous equations to solve (which we know how to
solve, with some difficulty!), we end up with a system of simultaneous differential equations. These are
beyond the scope of this course since they require rather advanced mathematics (specifically, you can
use something called Laplace transforms to recast your simultaneous differential equations as good old
regular simultaneous equations), although if time permits at the end of the course I can show you how
to set these up, and perhaps also how to solve them using numerical methods!
V.
Inductors and Inductance
Consider a coil of wire in the form of a solenoid. The coil has a
radius r (and cross-sectional area A = πr2), length l and N turns
(usually expressed as n = N/l turns per unit length) carrying
current I. From Ampere’s Law we calculate the magnetic field B
inside the solenoid as
B = µ0In
where µ0 = 4π x 10-7 T●m/A is the permeability of the vacuum. Furthermore, if this current I changes at
a rate dI/dt, the magnetic field will have a rate of change given by the time derivative of the above:
dB/dt = µ0n dI/dt
Finally, we know from Faraday’s Law that a changing magnetic flux generates an induced EMF given by
εind = -dΦB/dt
where ΦB is the magnetic flux, and the minus sign denotes Lenz’s Law in that the induced EMF tends to
oppose the change in magnetic flux. In the case of the solenoid, ΦB = B (πr2) N, so
εind = -dΦB/dt = (πr2) N dB/dt = (πr2) N (µ0N/l) dI/dt = (µ0AN2/l) dI/dt
→
εind = -L dI/dt
Where L = µ0AN2/l is known as the “self-inductance,” usually referred to as simply “inductance” of the
solenoid. L is measured in units of V●s/A, but it comes up frequently enough that L is given its own unit,
the Henry, with 1 H = 1 V●s/A.
As with the Farad, the Henry turns out to be a rather large unit, so most practical inductors are
measured in mH or µH.
To increase inductance, inductors can be made with a larger number of coils (increase N), larger coils
(increase A) or make coils with thinner wires (so that n can be increased), but this can only be done to a
certain extent, as all will raise the internal resistance of the wires, giving rise to unwanted effects.
Instead, a piece of ferromagnetic or ferromagnetic material may be inserted into the core of the
inductor, which increases the strength of the magnetic field by a large factor. In this case µ0 is replaced
by µ, where µ represents the relative permeability of the core material, and is may be thousands of
times higher than µ0.
In circuits, the behavior of the inductor is described by its fundamental relation, εind = -L dI/dt, although
in practice the minus sign is usually omitted, and we’re forced to remember that the induced voltage is
induced in a direction that tends to oppose the change in current.
Finally, we may consider the energy stored in an inductor. There are several ways to derive this, of
which I pick what I consider to be the simplest. If we start with an inductor with no current going
through it, and “ramp up” the current toward a value I, there must be some ramping of the current
dI/dt, against which the inductor applies an induced EMF εind = L dI/dt (where I’m neglecting the minus
sign). The power (i.e. the rate at which work is done) at any moment is given by P = εindI, and the total
work done is:
U = ∫P dt = ∫εindI dt = ∫L I (dI/dt) dt = (1/2) LI2
That is, an inductor with a current I flowing through its coils stores a quantity of energy U = (1/2) LI2 in the
magnetic field the current generates. As I increases, the field strength (and thus the energy stored in
the field) increases, and as I decreases, the energy stored in the field decreases. Thus, like the capacitor,
the inductor stores electrical energy, which may be returned to the circuit some time later.
VI.
The Simple RL Circuit
A simple RL circuit is depicted at right. Since all of the functional
dependences in RL circuits are very similar to what we saw with RC
circuits, I will give a rather abbreviated discussion of the inductor’s
behavior.
We will assume that for t < 0 the circuit has sat silently with I = 0.
Therefore, for all this time, the inductor has been happy, as dI/dt was also zero, so VL = εind = 0.
At t = 0, someone closes the switch. At that moment, we can write KVL as:
ε - VR – VL = 0
where VR = I R, and VL = L dI/dt (where we’ve omitted the minus sign).
Note that the current in the circuit cannot change discontinuously: if this were the case, dI/dt would be
infinite, and the induced voltage would be infinite! That is unphysical. (This is analogous to the
capacitor disallowing any discontinuous changes in voltages; the inductor does the same with current.)
For this reason, at t=0, no current may flow. I = 0, so VR = I R = 0, and it follows that VL = ε. Using that,
we can state that
dI/dt = VL/L = ε/L
is the rate at which the current begins to increase at t = 0. (In analogy to what we saw with capacitors,
the limit in which the current instantaneously goes from zero to its final value corresponds to L = 0.)
Therefore, as time goes on, the current in the circuit gradually increases. If I is gradually increasing, then
VR = I R must be as well. Then VL = ε - VR must be gradually decreasing, and thus dI/dt must decrease
with time (although I itself is increasing!).
After a very long time, the circuit reaches a steady-state condition, in which all the V’s and I’s are
constant in time. If this is the case, dI/dt = 0, so VL = 0, and it follows that VR = ε, so I = ε/R is the longterm steady-state current that eventually flows.
Therefore the rules we can establish are:
When the switch is initially closed at t=0 and there is a “shock” of current that tries to flow, the inductor
acts like an open circuit in that it does not permit any current to flow.
After a very long time, the circuit reaches a steady-state condition, and the inductor acts like a short
circuit.
What about intermediate times? Here we will write down our KVL equation again:
ε - VR –VL = ε - IR – L dI/dt = 0
We note again that this is an equation involving only I, its derivative dI/dt, and a few constant values, yet
miraculously this linear combination of I, dI/dt and a constant value somehow always adds up to zero!
This implies that I and dI/dt have some functional form such that a derivative of I returns something
with the exact same functional form. Again, we will “guess” that the solution is something exponential,
but given that the current starts from zero and increases towards its max value (rather than vice versa),
we’ll use the “other” exponential formula we saw with RC circuits, and guess that I(t) = If(1-exp(-t/τ)),
where τ is the familiar time constant (to be solved in terms of L and R), and If represents the eventual
steady-state current that flows after a long time. If I(t) = If(1-exp(-t/τ)), then dI/dt = +If/τ exp(-t/τ), and
ε - VR – VL = ε - IR – L dI/dt = ε - RIf(1-exp(-t/τ) – LIf/τ exp(-t/τ)
Which we can re-arrange to group all the exponential terms, and all the constant terms, together:
ε - RIf - exp(-t/τ) (RIf – LIf/τ) = 0
Now I apply a little “trick” – I now have two terms, one of which is constant (ε - RIf), and one of which is
has a complex time dependence: exp(-t/τ) (RIf – LIf/τ). Since both terms have different time
dependence, but miraculously subtract out to zero at all values of t, this tells us that the coefficient of
the “exponential” term (RIf – LIf/τ) must be zero (so that the exponential time dependence is multiplied
by a coefficient of zero), and separately, then, the constant term ε - RIf must be zero (so that the overall
equation balances out). Therefore
ε - RIf = 0
--and--
RIf – LIf/τ = R – L/τ = 0
from which I determine that If = ε/R (which matches what I found earlier by “shorting” out the inductor),
and
τ = L/R
which you can compare to the τ = RC which I found for RC circuits.
With this in hand, I can go back and write down equations for all my LR circuit parameters:
I(t) = If (1-exp(-t/τ))
with If = ε/R and τ = L/R; I increases from zero toward If as t increases
t
VR(t) = R I(t) = ε (1-exp(- /τ))
VL(t) = ε - VR(t) = ε - ε (1-exp(-t/τ)) = ε exp(-t/τ)
dI/dt = VL(t)/L = ε/L exp(-t/τ)
And we’re done with the simple LR circuit!
One note is that the equations describing the LR circuit are essentially the same as those describing RC
circuits; this comes about because the fundamental relation for capacitors, I = C dV/dt, has such a similar
form to the defining relationship for inductors, V = L dI/dt.
VII.
Dealing with combinations of inductors
At this point we know how to solve the “simple RL circuit.” Now, if we were given a more complex
circuit, in analogy with what we did with RC circuits, a natural idea would be to try to reduce the circuit
to a “simple RL circuit.”
We now know how to reduce series and parallel combinations of both resistors and capacitors; do
similar relationships hold for inductors? The answer (as you may expect), is yes.
If three inductors are in series, they all have the same current (and thus dI/dt) through them. Since they
are in series, the total voltage across the three inductors Vtot is equal to the sum of the three individual
voltages. We can therefore replace the three inductors with a single Leq which obeys the fundamental
relation for inductors:
Vtot = Leq dI/dt = L1 dI/dt + L2 dI/dt + L3 dI/dt = (L1+L2+L3) dI/dt
Thus, for inductors in series,
Leq = L1 + L2 + L3 + …
which is the same formula we found for resistors in series, and for capacitors in parallel.
For inductors in parallel, we know that all must have the same voltage V across them, which means that
a changing current dI/dt will distribute itself among the inductors as
V = L1 dI1/dt = L2 dI2/dt = L3 dI3/dt = …
where I = I1 + I2 + I3 + …, so dI/dt = dI1/dt + dI2/dt + dI3/dt + …
So we can replace the parallel inductors with a single inductor with Leq, such that
V = Leff dI/dt = Leq (dI1/dt + dI2/dt + dI3/dt + …) = Leq (V/L1 + V/L2 + V/L3 + …)
After canceling the common factor of V, and dividing both sides by Leq, we find:
1/Leq = 1/L1 + 1/L2 + 1/L3 + …
for parallel inductors, which mirrors the familiar “one over” formula we found for resistors in parallel,
and capacitors in series.
Therefore, if we have an RL circuit, we may try to reduce it by (1) reducing all the resistors into a single
equivalent resistance Req; and (2) separately reducing all the inductors into a single equivalent
inductance Leq. If that is not possible, we can try using our network theorems (Thevenin, Norton,
superposition) to replace a nasty RL network with things we can solve.
Failing that, we have to revert to our Kirchhoff-based methods, which as with RC circuits, will generally
result in a system of simultaneous differential equations, which are beyond the scope of what we will do
in this course.
VIII.
Practical aspects of capacitors and inductors
So far I’ve concentrated on circuits composed of “ideal” capacitors and inductors.
A. Ideal and Real Capacitors
The ideal capacitor was modeled as a pair of parallel plates. The plates themselves are completely
without resistance, and there is zero leakage of current between the plates. Real capacitor plates have a
finite amount of resistance (which effectively is in series with the capacitance), and real capacitors have
leakage between the plates (which is effectively represented as a leakage resistance across the
capacitor.
As I mentioned earlier, the Farad is an enormous unit. Typically, a reasonably-sized and spaced pair of
parallel plates will have a capacitance on the order of pF (you can check this for yourself by plugging in
“reasonable” values to C = ∈0A/d; since ∈0 ~ 10-11 F/m, you need to plug in very “unrealistic” values for A
or d -- orders of magnitude away from the familiar 1-ish values we know and love -- to get a capacitance
that differs much from the pF regime.
So let’s think about how capacitance is affected by the geometric factors in C = ∈0A/d. Note that if the
plate area A increases, the capacitance increases. So one way to make a capacitor with larger
capacitance is to increase the plate area (usually, this is done by building a capacitor out of multiple
“layers” of small plates, or by having large-area plates rolled into a cylinder, rather than having huge flat
plates that take up a large amount of space).
Capacitance can also be increased by making the plate separation d as small as possible. Though this is a
bit counterintuitive, it makes sense. For a given value of ±Q on plates of area A, the electric field E
between the plates is fixed at Q/∈0A. If the distance between the plates is made smaller, then the
potential difference V = E*d between the plates decreases. Hence, for a given potential V applied
between the plates, a larger charge Q can be stored, and C = Q/V is larger.
A third way to increase C is to somehow decrease the E field between the plates, for a given value of Q.
Unfortunately, E = σ/∈0 comes from fundamental laws of physics (in particular, ∈0 is set by the
properties of the vacuum in our universe), so we can’t change that if there is only vacuum between the
plates without traveling to another universe. (Good luck with that!)
However, many insulating materials exhibit dielectric properties. This means that while the electrons in
the material cannot travel freely to conduct electricity (which is why they’re insulating), the charges can
displace a small amount from their equilibrium position, giving rise to an electric field which tends to
“screen out” the applied field – the negative charges lean slightly toward the positive plate, and the
positive charges displace a small amount toward the negative plate, giving the illusion that the actual
charges on the plates are less than ±Q. But the only issue that matters to us at present is that when a
dielectric material is present between the plates, ∈0 in the above formulae is replaced by ∈ = κ∈0, where
κ is known as the relative permittivity of the material. The vacuum has κ = 1 by definition, while air has
κ = 1.006. Paper typically has κ = 2-4, while mica has κ=5, and typical glasses have κ = 5-10. Some
ceramic materials, however, have κ ranging from 100 to over 10,000, so these can give a substantial
increase in the value of capacitance!
B.
Classification of Capacitors
As done with resistors, we should come up with a scheme to classify capacitors. The following
categories are useful:
-
Fixed or variable value of capacitance?
Breakdown voltage
Value and tolerance of C
Type of dielectric material
While the vast majority of capacitors in practical use have fixed values of capacitances, there are some
cases where a changing value of capacitance is required. In these cases, a variable capacitor is used.
Typically these have an array of air-spaced plates, which can be rotated relative to each other to vary
the effective plate area A by turning a knob. Variable capacitors can also be made by having two plates
whose separation can be adjusted by turning a screw; these are typically referred to as trimmer or
padder capacitors. While the total capacitance is typically only in the pF range, these are crucial in
tuning circuits.
Another consideration regards the presence of high voltage between closely-separated plates. If the
voltage across an insulator (such as the dielectric between the capacitor’s plates) becomes high enough,
breakdown occurs. This means that whatever charges have accumulated on the plates will short across
the insulating space between them: the device ceases to act as a capacitor, and in fact may be damaged!
For this reason, all capacitors have a maximum voltage rating, just as resistors have a maximum power
rating. Some dielectric materials have extremely high breakdown voltages, even for small thicknesses,
so these types of capacitors are usable at high applied voltages.
Most typical capacitors in use have thin layers of materials such as mica or various ceramics as the
dielectric, with the plate area increased either by rolling large plates into a cylindrical form, or by
stacking layers of plates. Ceramic and mica capacitors can be made with capacitance values from a few
pF up to roughly 0.05 µF, but have the advantage of rather high breakdown voltages (typically at the
very least several hundred, and usually several thousand, volts).
Some very old electronic devices use capacitors with paper as the dielectric material. Very frequently
such paper capacitors were hand-rolled, and used due to ease of manufacture and low cost, although
nowadays ceramic and mica capacitors are much more common. Since paper capacitors can be made
very cheaply, very large (in size) units can be made, and with rather small plate separations d (due only
to the thickness of the paper), potentially may have capacitances up to about 1 µF, though typically they
are much smaller than that.
Capacitors can be made with thin films such as mylar or polystyrene as the dielectric material. The
advantage here is that d can be extremely small, so film capacitors can have C up to several µF. They
have an advantage in that their capacitance is usually less sensitive to temperature changes than other
types, but a disadvantage in that their breakdown voltages are somewhat smaller than for mica and
ceramic capacitors (typically 50-1000 V).
One special type of capacitor is made by repeated deposition of alternating layers of metal film and
insulating materials. These are called electrolytic capacitors, and because they can have very small
effective plate separations d, and large effective plate areas A, can have large capacitances, on the order
of tens to hundreds (and sometimes thousands) of µF. However, this comes at a tradeoff. Firstly, since
the dielectric layers are so thin, breakdown voltages are often relatively small compared to those in
more “standard” capacitors (typically on the order of tens of volts, though some are on the order of
hundreds, and others are < 10 V). Secondly, because of the way the layers are applied, electrolytic
capacitors can generally only handle applied voltage of one polarity, and are clearly marked with “+”
and/or “-“ terminals to indicate how they should be wired into a circuit. If they are wired with reverse
polarity, they can be damaged! Electrolytic capacitors also are prone to drying out as they age, so over
time they are one of the most common components that needs to be replaced.
A special type of electrolytic capacitor known as the tantalum capacitor is also widely used. These can
also be manufactured with very tight tolerances on capacitance values and are less susceptible to drying
out. For these reasons and others, they are widely used in computers and related electronics. Tantalum
capacitors typically have capacitance values on the order of 1-50 µF with breakdown voltages on the
order of tens of volts.
With this deposition method taken to extremely thin layers, so-called supercapacitors can be made.
These typically have extremely small limiting voltages (typically a few volts), but can have capacitances
in the range of F in a relatively small package!
Finally, being that I am so enamored of quaint old electronics terms, you’ll sometimes hear capacitors
referred to as condensers.
C.
Ideal and real inductors
Earlier, we modeled inductors as resistanceless coils of wire obeying Faraday’s Law. In reality, these
coils have a non-zero resistance, which is effectively in series with the inductor. There is also a small
amount of “parasitic” capacitance between neighboring wires in the coil (since each turn is at a slightly
different voltage than its neighbors), which is effectively in parallel with the inductance. Typically, larger
inductors (which have larger coils with more windings) are more seriously affected by these stray
resistances and capacitances. And in AC circuits, inductors become increasingly badly-behaved as
frequency increases.
In addition, there can be significant energy losses due to resistive heating of the core material; the
induced electric fields not only generate EMFs in the inductor’s coil, but in the core as well. Since the
core material has nonzero resistance, these eddy currents can result in significant energy losses,
particularly in cases where the applied current changes rapidly (as in high-frequency applications). For
this reason, insulating materials, or thin laminated layers of conducting material, are often used as
cores, in order to reduce the losses due to eddy currents.
Another practical concern arises from the construction of inductors. While resistors and capacitors (as
well as diodes and transistors) can very easily be miniaturized, and fabricated using methods conducive
to construction on integrated circuit boards, inductors (like the vacuum tubes of old) actually do need to
be rather large in order to get a sizable inductance, and made in a particular physical way (due to the
need to generate a sizable magnetic field to induce back EMF) which does not lend itself readily to
integrated circuit construction. For this reason, inductors have become somewhat of a “fifth wheel” in
electronics. They’re inconvenient to deal with, so electronics engineers will typically look for ways to
avoid using inductors in circuits, usually using combinations of capacitors and active devices (usually
operational amplifiers, which are constructed from combinations of transistors). For this reason,
inductors are not used nearly as often as most of the other components we’ll study. That said, there are
some applications for which they are crucial, such as radio-frequency tuning applications, high-power
voltage transformers, and of course the Taser. So unlike the vacuum tube, whose applications were
nearly entirely supplanted by the vastly more desirable transistor, inductors have a long future of use in
their “killer apps,” although they are far less common than resistors, capacitors and the solid-state
devices we’ll study later in the semester.
Finally, we know that a changing current through an inductor induces a changing magnetic field in its
core. Some of that changing magnetic field escapes from the core, and can impinge on other inductors
that happen to be in the vicinity (even if they’re part of a separate circuit!), so that a changing current in
one inductor can actually induce a voltage in another! This is known as mutual inductance. In some
cases, mutual inductance is desired; the prototypical example of this is in a transformer. In other cases,
it is unwanted and unexpected, in which case it leads to the circuit behaving in undesirable ways. So
care must be taken to minimize these stray fields, and to ensure that the stray fields that do escape have
minimal effects on other inductors. And because these stray induced voltages are a function of
geometry and other aspects that are incidental to circuit design, this effect is generally impossible to
model from a schematic diagram. In other words, very bad.
D. Classification of inductors
Just as we did with resistors and capacitors, we can come up with some schemes to classify inductors.
However, inductors as a group tend to be a bit more homogeneous, in that they’re all more or less the
same: a coil of wire. But we can come up with a few classification schemes:
-
Fixed or variable?
Geometry and core material
Value of inductance
Current/voltage ratings
Just as with resistors and capacitors, most inductors you’ll run across have fixed values of inductance.
But on occasion a variable-value inductor is required (most often, in a tuning circuit). The typical way
this is accomplished is with an adjustable core. A ferromagnetic or ferromagnetic core material
substantially increases the magnetic field inside the coil, so an inductor with a core that can be partially
inserted or removed, so that some of the coils are wrapped around a magnetic core, and some enclose
only air, can have a widely-varying total effective inductance.
There are two principal geometries in use for inductors. There is the cylindrical solenoid shape, which
we considered in the above section on inductors. These are the simplest to construct, and are also the
most readily amenable to simple calculations in “textbook” problems. But toroidal, or donut-shaped,
inductors are common as well, in which a ring-shaped core is wrapped with wire. These fundamentally
work the same way as solenoid inductors, but with one significant advantage: since the core material is
in the shape of a ring, the “stray” fields escaping the core are much smaller, and the effects on other
inductors in the circuit should hopefully be smaller.
Like the Farad in capacitance, the Henry is a rather large unit. Most real inductors have inductance
values on the order of µH to mH. Typically, the inductance value, and tolerance, and any appropriate
limits on applied voltage or current, are stamped on the body of the inductor.
Finally, in continuing my tradition of mentioning old electronics terms, old-timers (who, sadly, are
becoming fewer in number as the years go by) will frequently refer to inductors as “coils” (should be
obvious why) or “chokes” (due to their effect of “choking” rapid changes of current).
Download