Some notes on the Poisson distribution Ernie Croot October 7, 2010 1 Introduction The Poisson distribution is one of the most important that we will encounter in this course – it is right up there with the normal distribution. Yet, because of time limitations, and due to the fact that its true applications are quite technical for a course such as ours, we will not have the time to discuss them. The purpose of this note is not to address these applications, but rather to provide source material for some things I presented in my lectures. 2 Basic properties Recall that X is a Poisson random variable with parameter λ if it takes on the values 0, 1, 2, ... according to the probability distribution p(x) = P (X = x) = e−λ λx . x! By convention, 0! = 1. Let us verify that this is indeed a legal probability density function (or “mass function” as your book likes to say) by showing that the sum of p(n) over all n ≥ 0, is 1. We have ∞ X p(n) = e −λ ∞ X λn n=0 n=0 n! . This last sum clearly is the power series formula for eλ , so ∞ X p(n) = e−λ eλ = 1, n=0 1 as claimed. A basic fact about the Poisson random variable X (actually, two facts in one) is as follows: Fact. E(X) = V (X) = λ. Let us just prove here that E(X) = λ: As with showing that p(x) is a legal pdf, this is a simple exercise in series manipulation. We begin by noting that ∞ ∞ X X e−λ λn E(X) = np(n) = n· . n! n=0 n=1 Noting that n! = n · (n − 1) · (n − 2) · · · 3 · 2 · 1, we see that we may cancel the n with the n! to leave an (n − 1)! in the denominator. So, pulling out the e−λ and cancelling the n, we deduce that E(X) = e −λ ∞ X n=1 ∞ X λn λn−1 −λ = λe . (n − 1)! (n − 1)! n=1 (Note here that the term n = 1 has a 0! in the denominator.) Upon setting m = n − 1, we find that E(X) = λe −λ ∞ X λm = λe−λ eλ = λ, m! m=0 and we are done. 3 A sum property of Poisson random variables Here we will show that if Y and Z are independent Poisson random variables with parameters λ1 and λ2 , respectively, then Y +Z has a Poisson distribution with parameter λ1 + λ2 . 2 Before we even begin showing this, let us recall what it means for two discrete random variables to be “independent”: We say that Y and Z are independent if P (Y = y, Z = z) = P (Y = y and Z = z) = P (Y = y)P (Z = z). This is not really the “official” definition of independent random variables, but it is good enough for us. Letting now X = Y + Z, we have that P (X = x) = P (Y + Z = x) = X P (Y = y, Z = z) X P (Y = y)P (Z = z) y,z≥0 y+z=x = y,z≥0 y+z=x X e−λ1 λy e−λ2 λz 1 2 · y! z! y,z≥0 = y+z=x −(λ1 +λ2 ) = e X λy λz 1 2 . y!z! y,z≥0 y+z=x Now we require a little trick to finish our proof. We begin by recognizing that the y!z! is the denominator of x! x x . = = z y!z! y So, if we insert a factor x! in the sum, then we will have some binomial coefficient. Indeed, we have e−(λ1 +λ2 ) X x y z λ λ. P (X = x) = y 1 2 x! y,z≥0 y+z=x Notice that from the binomial theorem we deduce that this last sum is just (λ1 + λ2 )x . 3 So, e−(λ1 +λ2 ) (λ1 + λ2 )x , x! and therefore X has a Poisson distribution with parameter λ1 +λ2 , as claimed. P (X = x) = 4 An application It turns out that prime numbers in short intervals obey a “Poisson distribution”. To discuss this, I need to tell you about the Prime Number Theorem: Prime Number Theorem. The number of primes p ≤ x has size about x/ log x, where here the log is to the base-e. That is, if π(x) denotes this number of primes, then π(x) = 1. x→∞ x/ log x π(x) ∼ x/ log x, which means that lim In what follows, we adopt the notation π(I) to denote the number of primes in a given interval I. Now suppose that x is a “really large” number, say x = 10100 . Then, if you pick a number n ≤ x at random, there is about a 1/100 log(10) (about 0.43 percent) chance that it will be prime; and we expect that a typical interval [n, n + log n] ⊂ [1, x] will contain about one prime. In fact, much more is true: If you pick n ≤ x at random, and choose λ > 0 and j not too large (say λ, j ≤ 20), then the number of primes in [n, n + λ log n] roughly obeys a Poisson distribution: P(π([n, n + λ log n]) = j) ≈ e−λ λj . j! Notice that I didn’t use equality here; in order to get equality we would have to let x → ∞ in some way – certainly the larger we take x to be, the closer the above probability comes to e−λ λj /j!. Given what you know about Poisson distributions, it is an interesting exercise to figure out why one would expect the primes to be “Poissondistributed” as stated above. 4