Chapter 1 Basic Probability Theory Dep An outcome specific result a is Def The sample space S Def An event A is is the set of all possible outcomes subset of outcomes a Def The events Ai Az in S partition of S if they mutually disjoint and their union is S Theorem If are RBs Bz is a a partition of S then for any event A YILANBi A S B are Bz B A AABst An Bat AnBz g Def A P function called which assigns a numerical value to events is probabilitypunction if it satisfies the Axioms of a Probability 1 PLA P s 2 3 P is O L If AsAz a PCAUB function are disjoint PffA from the space PlA PCB if A n B are disjoint ÉPLA of events to the real line Theorem Forevents A and B 1 PLAY I PCA 2 PCO O 3 PCA El 4 If p then PLA s PCB ACB 5 PLAUB 7 PLANB Z If likely Dep If PCB PCB PAIT PCB an experiment symmetric the in sense outcomes ai an that each outcome which are is equally IN P ai then L has N the conditional probability of A given B is 0 PLANB PCB PCAIB Dep PLANB PCA t PCB 61 PLAUB EP A Theorem probability Largersets have larger The events A and B are independent A and B have absolutely outcome of one has no no effect influence on if PCANB PCA PCB on each other the outcome the of the other Theorem If A and B then P AIB are independent with PCA PCA PCBIA PCB O and PCB o Theorem If Ba Ba then Theorem If is 0 of S and P Bi PCB a so P BIA P A P AIB P P BIA PCA PCBAc B of Def A collection O for all i É PLAIBIP Bi E P AMBI PLA PLA a partition sets is a pay sigma BIA PCA PCB field if it satisfies the following 3 properties HOEB 2 3 EB then A EB If As Az EB then QiAi EB If A Def The Borel Sigma field is the smallest sigma field on IR It contains all open intervals containing all open intervals at and closed intervals and their countable unions intersections and A of complements be generated from a finite collection by taking all unions intersections and sigmafield can events complements Chapter 2 Random Variables Dep A random variable is the sample real valued a S space outcome a ruction from to the real line IR Represent random outcomes numerically Think of value is number or Dep Dep X a random variable and unknown t is discrete if infinite number elements X there is is a with this realization a as a specific outcome The set it If as a random object whose of a discrete set it has It a finite such that or countably PEX et I then discrete random variable The smallest set It property is I The support is the positive probability Jt 3Taitz of try X the support range of outcomes which of receive occurrence support points Def The probability mass function of a random variable is ITA PK the probability thattherandom variable X equals a specific value x When evaluated at the suppoint points 4Function that gives the IT we write IT IT es probability that a discrete random variable is exactly equal to some value x If X is Theorem g It function random Def For the V g X for G C IR then Y is also expectation random variable X of series X a is X with 25 support is TT ETH While some variable a discrete if the and a random variable non convergent Etr is either infinite or notdefined is conveyed random If Ete is non random since itis a fixed feature of the distribution Theorem Proof For any Etat GH Etatbe IE at GEEK at bij IT FEET a a t Def The and b constant a b EEN distribution function CDF is a discrete random variable with FCT Cumulative sum of É TE the PLA E x the 3143 probability of the event For FL support probabilities less points than j I The CDF is constant between the support points faction with jumps at each support point Theorem If FA is CDF a s FG is 2 EM Feel 0 3 fits Fk L 4 Feel is right continuous At non decreasing right since If Xu Fk meaning fix FG Fk points where discontinuous Def step and FK has a step Feel is to the left but continuous t the FL Fk PIKE is continuous x then X is continuous random variable PH EM PEX x The probability that equal PUA to a big Fete X FA o a continuous random variable is be specific value must PLAS X PAIN PH exte FK L FK zero Def When FLA is differentiable its density Theorem A function fix a is density function PDF is fee if and only if I FCA 20 HX Ifield 2 Plaek b L Jab field The probability that X is in the interval ta b is the integral of the density over ta b This shows that the area underneath the density functions are probabilities Def The support It of a continuous random smallest set containing Reifel x Def If X is continuously distributed expectation is defined Def the is the with density feel its as Exfelde EFX when variable integral is convergent The mean of X The variance of X is is M ETX EE var X ETH ER 8220 X is degenerate ie X if for some is non random r 52 C 0 Pk c L dad The Theorem var ETH X For any random variable is X if get EIGHT g EEA function of an than the expectation of EFX then a concave function convex is a convex SEEK g ETA 4 A Vr o EEN function then If get is 62bark Theorem Var lat bl Theorem X standard deviation of expectation the is less transformation E ETXT Etlog KDE log EET exp EEN SETUP XD The moments of a the powers of X distributions are Dep The Mth moment of Def For Mm m I moment of X ETH ETHIm moments finite infinite finite or infinite values of ELI is M'm the mth central Odd moments Even X the expected or undefined is Def The moment generating function M The MGF is either finite or 4 Non negative Theorem If MIA finite for is O then D MIO L in OYE to iii 5Mt JE Itm in The the curvature of distribution or X is Jeep tx fade Etexplex t MOF to to Mct at of X t in a infinite neighborhood of ELF E x2 ETXm for any moment which is finite to encodes all moments of Chapter 4 Multivariate Distributions Dep A pair of bivariate random variables is numerical outcomes a function pair a of from the sample space to IR Dep A pair of bivariate random variables variables with a random distribution joint The joint distributionfunction of X Y is F x g PIX ex YE y P HEA n 38 93 F 4 weakly increasing in each argument and satisfies O E F x y 1 is Satisfies the following Plack sb C Y Ed Flak Y d F bd b FaC c Fla d F la c X F 6c C A pair a Fb F bd a Dep are two of random variables is discrete if there is discrete set 94112 such that Y ES L I PIG support of TEsTE XY and Rt's TY Dep The joint probability mass function ITCHY PEX x Def When F xy is Y y and differentiable its continuous density equals The b Flay Roy probability that the region in 1122 equals this Dep The Jfk Y ed c marginal in under the density area distribution PAIN of X PIX In the is ex YEA Fay lying continuous case Fx k lying I If uu The marginal densities of are density f x y the In dedy pair X Y lies random the g region Fx K Dep joint O fley L Plas is II fix practice we g dy treat du du X and Y Fy a y marginal Iff given curldudu a joint fay de PDF the same a over Def The expectation of real valued g K Y Efg X Y ay Ernay for the discrete case for the continuous case gk go y f I Ight X y de dy fay dxdy discrete distribution has a distribution function of Y given Fy lyin PLYEgl X For any e such Pl X that 4 Distribution function Def If Fuk lyk PLAe x given o X X on one variable Ig I Ix fee y dedy It E TX If x only depers E Ight Dep and II If 949 TWD o Efg X Y Note is x for the e e E x de the conditional K x is x o subpopulation where tix is differentiable withrespect to y and then the conditional density function of is FK y x de OFK GK Oy Y Density Def For Y for the of X and Y the continuous given X X is Fk for any subpopulation where x let A Hex tix conditional density or Y FS y x FK such that Fx t so B 31 193 The probability that they both occur is PLANB PLA e x Y Ey Flag The product of their probabilities is PCA and PIX Ex PLYEy PIB Def The random variables if for all k n are statistically independent Fy x Ffs probability is equal to the product of Def The discrete random variables X and if for all k n they Y are probabilities statistically independent y TK g Ip X and Y s y FG g The joint X and Y EG E Tx have are statistically x Ty y differentiable distribution function independent if for all k n y a x Theorem If X and Y are independent and continuously distributed f hey Fuld x FXIX K ly If X and Y FX x independent are the conditional density the marginal density i.e he does not affect the shape of the density of Y equals Theorem Bayes Theorem for Densities Fxx Fy k SK Theorem If X and Y rly FA x are fy y independent g IR IR and hi IR IR Eth Y Loo such EISK HH Def If X and Y have finite then for any that Efg X functions 200 and EIGHTEEN Y variances X and Y is Cov X Y ELK ELKIN EEN EEN EEN EEN the covariance between Def If X and Y have K and Y is finite the variances correlation between Cork Y Corr X Y varlet var Y Theorem X and Y are independent with finite variances then ETH ETH ETH and hence Cov X Y 0 Theorem If X and Y have If bark var Ktv Theorem For any variances finite var Y random variable X 2 Cov X Y and Y EVEENELT ETH Cauchy Schwarz inequality Dep The conditional expectation of expected value of the and is written as For discrete variables ELY he For continuous Y ELY Xx Ygiven Xx conditional distribution ETUI text me the is fuk lyk Y within Expected value on FEI Tx the subpop The population f which tee e Expectedvalue of g for fun Y within the infinitesimally small for which ye dy tax t Y flat dy Iffhs dy Remember Et VI X x is not random it is a feature of the joint distribution However Et VIX is a random variable a transformation of X In other words EE YIN is a function of the random variable X JEENA FAH de ELENA II Iffy Y FK lyk Exceldyde flay dydx E EY The average Theorem If ELY across Loo group averages is the grand average then ELETTIN EET conditional variance of Y given X x is the variance of the conditional distribution Fuk six and is written as var Y l x or 8241 It equals Def The var HK ETH ELINT ELVIRA Theorem vartl LELY Xx KIK EEYHtvarg.IE variance variance Dep A multivariate random vector is a space Dep to 112m written The joint distribution F PA x Def When FG is Xm Xm Xm the joint probability vectors mass P Xx IT e is ki ka sample function is PCIE Xs ex Def For discrete random function X as from the function continuous and differentiable its joint density feel equals F Def The expectation of its OMFG X ON of r X EIR OXm is the vector or expectations elements Eth EEK EFX Il Def The mem covariance varTX z matrix of Xt IR is ETH ETH X EEN I EE i i É One i one Z Theorem Any mem covariance matrix I Symmetric 2 Theorem Positive Z satisfies Z For any semi definite If X EIR with I Mtt ato a mass expectation M and mem covariance matrix I and A is gem then AX is a random vector with mean AM and covariance matrix AZA Dep The joint distribution function of X HE IRMARm FG g P XS x Y Ey Dep The joint density function of X HE IRMARm f x y datmy F Fx density functions Fy Def The conditional densities of Y y is OMy Jfk yldy x is xy DX Def The marginal of X and Y are fay de y Y given X x and X given are FK SK Def za z o Iff finely III The conditional expectation of Y given Xx ELY X X Syfy yle dy is Chapter 7 Law of Large Numbers Large sample asymptotic approximations Def A n has an sequence ou lying an oo a limit a Example for 0 tanks to 0 and 1 Fix 820 2 Find As such that Ian 3 If 2 can be done for converges to 0 n Let 8 If an a as there is some ng Coo n An as a Converges Ovitt solving written od to show an found by setting oo Ian al ES Mm Proof a if for all 8 such that for all n ng an as n 12 n 0 Is 8 In then Ian I S e s A ES n 8 for ne ng any 8 then an Def A sequence probability for all We call 8 c of random variables Zn e IR converges in to c as n soo denoted Play Zn c if 0 big PII Zn c hit Pt 4 s Zn L E8 or o the plim of Zn of random variables concentrating about a point For any Sso the event 312n c 283 Occurs when Zu is within 8 of C Pt Zn c les is the probability Sequence Pt Zn CI es approaches I as n Note The definition concerns the distribution variable not the realizations oo of the random Note The definition uses the concept of a conventional limit but applied to a sequence of probabilities not directly to Zn or their realizations Theorem For any random Chebyshev variable X and any 8 0 EtK EtX py.fm zgyevarty 82 4 The probability that the randomvariable differs fromitsmean by with finite mean me and finite variant Take any X PIK M1 We want to bond the probability distributions and all 8 Theorem For any sequence Vartzn 8 for all of random variables Zn such that Zn EEZn then O P O as N soo 0 oh is sufficient Chebyshev's for Inequality Theorem For any random variable X and any 8 Markov ETH1141412837 Et pix Theorem WUN gy ti If EIA 8 independent and identically distributed are then L OO Samplemeant sample expectation Proof Since mean approach the can PC converges 4 For any distribution Xi is use H M as I I Xi An The 0 a the E n 00 this equivalent to P ELK fit P o in probability to the population with a finite expectation thesample mean population expectation random variable in probability as with Chebyshev Inequality s Itn Mlse and Varlet where nooo ETH we Eiti In PL IE MI as n Varki NTI E o 850 of a parameter is consistent if 00 There is the f Eez E E É An estimator Dep Var In size n sufficiently large such that sample a I will estimator be arbitrarily close to the true 0 value with high probability Dep A function Ht CH E S A function hat is continuous that hall is continuous if at te if te o 78 o E E small changes in an input result the output small changes in r hidthe 1 I 414 ha E W e's Theorem CMT If Zn then Ig e c as h Zn P II Zn ell E E n P soo hic I and he is as n soo as n soo I P Ilh Zn h C 11 EE I as nooo continuous at C in Proof Let Eso hf is continuous at C that halls E 78 It 41 8 Since Evaluated at Zn x we Pliant hulled Zak Since This implies Theorem Let then C nan find PtyZn ell Pll Zn ell P se I se hic I denote the set of distributions for which ETH COO Then for any sample size n and any E so I For P PTIta ETH e L there is a probability distribution such that the probability that the sample mean differs from the any n population expectation by E is Failure one of Uniform convergence while ETA it does not converge L is for which ETA For any sample the sample mean In converges to uniformly across distributions size there is a distribution such that is not close to the population expectation Theorem I Let EB Var X N s distributions which satisfy of o as PIItn EEN O e A sufficient condition is to found a moment assume vark EB for some Boo greater than one this restriction the WUN across the set of Define 00 for all E o Xi kilo as n distributions and of surely to c random variables as Pfling Za It of 4 In computes a the c sic O ZuelR Zn converges c almos in L probability of a limit rather than probability a which tot In Eiti Then e denoted noo for oo FI PIIFT EEP sequence holds uniformly distributions Let It denote the set of 82 Def A Then for all E 00 00 Under Theorem B for some SEE 4 set denote the implies Zu P C the limit Theorem SUN If Xi ETA are then COO and identically distributed and independent noo as In t.E.fi a s Efx Chapter 8 Central limit Theory Def let Zn be a vectors of random variables or We that PEZ sequence with distribution Gn u need say Zn converges in distribution to Z as n soo denoted d Z Zn if for all u at which Glu PEZ u is continuous Gn A sequence of the sequence a write the equivalent WUN From the Since convergence in convergence To In to obtain Gulu we s know converges to is degenerate 1 d c which is Zn as PEZ 1 c that I s M probability to a constant is the same this in Glu convergence Zn in distribution a converges in distribution of distribution functions the limit distribution G we can n soo as Zn random variable distribution auction limit When Glu u means non degenerate distribution as In dm we need to rescale rn F M Itn q Zn ETZn Moments of m which means Vartzn O Zn as O Etzi 82 ELZA Kafir E IZ Yn 304 E Ksdk Efta As the of the Kofi 82 O 10km 384 h O 115Ky kit lousy 1582 155 sample increases these moments converge to those normal distribution If X has finite rth moment then EEZ where Zn No 04 a to EEZ d 02 n soo Elza Zn Ern Fyi Theorem Zn Theorem For random variable with Z if Eteeplezn Mn t Eterpltzn Etelp EZ a finite exp MGF E converges for every te IR If ti Theorem CLT d Tn ta gu where The simple M ETA ETA and i id are and Loo then as nooo N 0,04 82 ETH M of averaging induces normality The distribution of the random variable rn ti p approximately process the same as NCO 02 when is u is large Tna Nlm For ZuelR multivariate Theorem Zn d Z ipp t Zn d t Z for every HEIR with L Let Xi a vectorvalued observations Varix EFX M let Zn rn In M Theorem If Xi ER are and In sample averages n z i id rn In pi and d ETH NCO E Coo then as nooo Theorem CMT If Zn d and h km Z as noo IR has the set of discontinuity points Dh such that Pl Ze Dn then h Zn d h z as nooo If his Convergence d then h zu continuous h Z in distribution is preserved by continuous transformations Theorem Slutsky If 3 Delta Method d Z and Cn d I Zn t Cn 2 Theorem Zn Zn Cn d d In If F E oo ther Ze d differentiable in n n as Zte E o se Cto if 5 and heal is continuously neighborhood of a other oo Tn Where H u ha ha Juhu H and H H In particular if bunco rn ha d ha V d o then as noo NCO H'VA as 0