Probability Distributions

advertisement
Probability Distributions
Signals and Systems in Biology
Kushal Shah @ EE, IIT Delhi
Random Variable
I
A number assigned to every outcome of an experiment.
I
A function whose domain is the set of all experimental outcomes
X :Ω→R
I
I
X is defined on probability space
I What is P (X = x )?
.
I Ex : For a roll of a fair dice, P (X = 5) = 1 6 = P (X = 1)
I 0 ≤ P (X = x ) ≤ 1
∀x
I ∑ P (x ) = 1
x
Ω can be a discrete or continuous set
Random Variable
I
A number assigned to every outcome of an experiment.
I
A function whose domain is the set of all experimental outcomes
X :Ω→R
I
I
X is defined on probability space
I What is P (X = x )?
.
I Ex : For a roll of a fair dice, P (X = 5) = 1 6 = P (X = 1)
I 0 ≤ P (X = x ) ≤ 1
∀x
I ∑ P (x ) = 1
x
Ω can be a discrete or continuous set
Random Variable
I
A number assigned to every outcome of an experiment.
I
A function whose domain is the set of all experimental outcomes
X :Ω→R
I
I
X is defined on probability space
I What is P (X = x )?
.
I Ex : For a roll of a fair dice, P (X = 5) = 1 6 = P (X = 1)
I 0 ≤ P (X = x ) ≤ 1
∀x
I ∑ P (x ) = 1
x
Ω can be a discrete or continuous set
Random Variable
I
A number assigned to every outcome of an experiment.
I
A function whose domain is the set of all experimental outcomes
X :Ω→R
I
I
X is defined on probability space
I What is P (X = x )?
.
I Ex : For a roll of a fair dice, P (X = 5) = 1 6 = P (X = 1)
I 0 ≤ P (X = x ) ≤ 1
∀x
I ∑ P (x ) = 1
x
Ω can be a discrete or continuous set
Random Variable
I
A number assigned to every outcome of an experiment.
I
A function whose domain is the set of all experimental outcomes
X :Ω→R
I
I
X is defined on probability space
I What is P (X = x )?
.
I Ex : For a roll of a fair dice, P (X = 5) = 1 6 = P (X = 1)
I 0 ≤ P (X = x ) ≤ 1
∀x
I ∑ P (x ) = 1
x
Ω can be a discrete or continuous set
Random Variable
I
A number assigned to every outcome of an experiment.
I
A function whose domain is the set of all experimental outcomes
X :Ω→R
I
I
X is defined on probability space
I What is P (X = x )?
.
I Ex : For a roll of a fair dice, P (X = 5) = 1 6 = P (X = 1)
I 0 ≤ P (X = x ) ≤ 1
∀x
I ∑ P (x ) = 1
x
Ω can be a discrete or continuous set
Random Variable
I
A number assigned to every outcome of an experiment.
I
A function whose domain is the set of all experimental outcomes
X :Ω→R
I
I
X is defined on probability space
I What is P (X = x )?
.
I Ex : For a roll of a fair dice, P (X = 5) = 1 6 = P (X = 1)
I 0 ≤ P (X = x ) ≤ 1
∀x
I ∑ P (x ) = 1
x
Ω can be a discrete or continuous set
Random Variable
I
A number assigned to every outcome of an experiment.
I
A function whose domain is the set of all experimental outcomes
X :Ω→R
I
I
X is defined on probability space
I What is P (X = x )?
.
I Ex : For a roll of a fair dice, P (X = 5) = 1 6 = P (X = 1)
I 0 ≤ P (X = x ) ≤ 1
∀x
I ∑ P (x ) = 1
x
Ω can be a discrete or continuous set
Probability Distribution Function
When Ω is a continuous set,
I P (X = x ) = 0 ∀x (usually)
I P (x < X < x + dx ) = f (x ) dx

Probability Distribution Function [PDF]

I f (x ) ≡
Probability Density Function


Probability Function
I Cumulative Distribution Function [CDF] or Mass Function
FX (x ) = P (X ≤ x )
ˆ x
=
−∞
fX (x ) dx
d
lim F (x ) = 1
x →∞ X
F (x )
⇒ fX (x ) =
dx X
ˆ ∞
f (x ) dx = 1
−∞
f (x ) ≥ 0∀x
Probability Distribution Function
When Ω is a continuous set,
I P (X = x ) = 0 ∀x (usually)
I P (x < X < x + dx ) = f (x ) dx

Probability Distribution Function [PDF]

I f (x ) ≡
Probability Density Function


Probability Function
I Cumulative Distribution Function [CDF] or Mass Function
FX (x ) = P (X ≤ x )
ˆ x
=
−∞
fX (x ) dx
d
lim F (x ) = 1
x →∞ X
F (x )
⇒ fX (x ) =
dx X
ˆ ∞
f (x ) dx = 1
−∞
f (x ) ≥ 0∀x
Probability Distribution Function
When Ω is a continuous set,
I P (X = x ) = 0 ∀x (usually)
I P (x < X < x + dx ) = f (x ) dx

Probability Distribution Function [PDF]

I f (x ) ≡
Probability Density Function


Probability Function
I Cumulative Distribution Function [CDF] or Mass Function
FX (x ) = P (X ≤ x )
ˆ x
=
−∞
fX (x ) dx
d
lim F (x ) = 1
x →∞ X
F (x )
⇒ fX (x ) =
dx X
ˆ ∞
f (x ) dx = 1
−∞
f (x ) ≥ 0∀x
Probability Distribution Function
When Ω is a continuous set,
I P (X = x ) = 0 ∀x (usually)
I P (x < X < x + dx ) = f (x ) dx

Probability Distribution Function [PDF]

I f (x ) ≡
Probability Density Function


Probability Function
I Cumulative Distribution Function [CDF] or Mass Function
FX (x ) = P (X ≤ x )
ˆ x
=
−∞
fX (x ) dx
d
lim F (x ) = 1
x →∞ X
F (x )
⇒ fX (x ) =
dx X
ˆ ∞
f (x ) dx = 1
−∞
f (x ) ≥ 0∀x
Probability Distribution Function
When Ω is a continuous set,
I P (X = x ) = 0 ∀x (usually)
I P (x < X < x + dx ) = f (x ) dx


Probability Distribution Function [PDF]
I f (x ) ≡
Probability Density Function


Probability Function
I Cumulative Distribution Function [CDF] or Mass Function
FX (x ) = P (X ≤ x )
ˆ x
=
−∞
⇒ fX (x ) =
lim F (x ) = 1
x →∞ X
ˆ
fX (x ) dx
d
F (x )
dx X
∞
−∞
f (x ) dx = 1
f (x ) ≥ 0∀x
Probability Distribution Function
When Ω is a continuous set,
I P (X = x ) = 0 ∀x (usually)
I P (x < X < x + dx ) = f (x ) dx


Probability Distribution Function [PDF]
I f (x ) ≡
Probability Density Function


Probability Function
I Cumulative Distribution Function [CDF] or Mass Function
FX (x ) = P (X ≤ x )
ˆ x
=
−∞
fX (x ) dx
d
lim F (x ) = 1
x →∞ X
F (x )
⇒ fX (x ) =
dx X
ˆ ∞
f (x ) dx = 1
−∞
f (x ) ≥ 0∀x
PDF and CDF
When Ω is a continuous set,
I
I
I
P (X = x ) = 0 ∀x (usually)
P (x < X < x + dx ) = f (x ) dx
´x
FX (x ) = P (X ≤ x ) = −∞
fX (x ) dx
Example of PDF : Schrodinger’s equation
ψ (x ) : Quantum Wave Function
−
h̄2 ∂ 2 ψ
= [E − V (x )] ψ
2m ∂ x 2
m : Mass of the particle
E : Energy of the particle
V (x ) : Potential energy due to the force field
|ψ (x )|2 dx : probability of a particle to be in the region (x , x + dx )
Discrete Distributions
Bernoulli Distribution :
P (X = 1) = p
P (X = 0 ) = q = 1 − p
Binomial Distribution :
n
P (Y = k ) =
p k q n−k ,
k
Ex : No. of forward steps in a random walk
k = 0, 1, 2, ..., n
Discrete Distributions
Bernoulli Distribution :
P (X = 1) = p
P (X = 0 ) = q = 1 − p
Binomial Distribution :
n
P (Y = k ) =
p k q n−k ,
k
Ex : No. of forward steps in a random walk
k = 0, 1, 2, ..., n
Discrete Distributions
Bernoulli Distribution :
P (X = 1) = p
P (X = 0 ) = q = 1 − p
Binomial Distribution :
n
P (Y = k ) =
p k q n−k ,
k
Ex : No. of forward steps in a random walk
k = 0, 1, 2, ..., n
Normal (Gaussian) Distribution
f (x ) = √
1
2πσ
e
2
−(x −µ)2
.
2σ 2
µ : Mean
σ 2 : Variance
I Distribution of velocities of molecules
I Central Limit Theorem (CLT)
∼ N µ, σ 2
Normal (Gaussian) Distribution
f (x ) = √
1
2πσ
e
2
−(x −µ)2
.
2σ 2
µ : Mean
σ 2 : Variance
I Distribution of velocities of molecules
I Central Limit Theorem (CLT)
∼ N µ, σ 2
Normal (Gaussian) Distribution
f (x ) = √
1
2πσ
e
2
−(x −µ)2
.
2σ 2
µ : Mean
σ 2 : Variance
I Distribution of velocities of molecules
I Central Limit Theorem (CLT)
∼ N µ, σ 2
Central Limit Theorem (CLT) : Bernoulli Trials
Xi ∈ {0, 1}
P {Xi = 1} = p
P {Xi = 0} = q = 1 − p
2
µ = p
σ = pq
X1 + X2 + · · · + Xn
Fraction of successes in n trials
Xn =
n
k
P Xn =
n
= b (n, p , k ) =
n
k
p k q n−k
By CLT,
f Xn = x
(x − µ)2
∼ N µ,
=q
exp
−
2 n
n
2
σ
2
2πσ n
"
#
1
(x − p )2
= q
exp − 2pq n
2π pq n
σ2
1
"
#
Central Limit Theorem (CLT) : Bernoulli Trials
Xi ∈ {0, 1}
P {Xi = 1} = p
P {Xi = 0} = q = 1 − p
2
µ = p
σ = pq
X1 + X2 + · · · + Xn
Fraction of successes in n trials
Xn =
n
k
P Xn =
n
= b (n, p , k ) =
n
k
p k q n−k
By CLT,
f Xn = x
(x − µ)2
∼ N µ,
=q
exp
−
2 n
n
2
σ
2
2πσ n
"
#
1
(x − p )2
= q
exp − 2pq n
2π pq n
σ2
1
"
#
Central Limit Theorem (CLT) : Bernoulli Trials
Xi ∈ {0, 1}
P {Xi = 1} = p
P {Xi = 0} = q = 1 − p
2
µ = p
σ = pq
X1 + X2 + · · · + Xn
Fraction of successes in n trials
Xn =
n
k
P Xn =
n
= b (n, p , k ) =
n
k
p k q n−k
By CLT,
f Xn = x
(x − µ)2
∼ N µ,
=q
exp
−
2 n
n
2
σ
2
2πσ n
"
#
1
(x − p )2
= q
exp − 2pq n
2π pq n
σ2
1
"
#
Central Limit Theorem (CLT) : Bernoulli Trials
Xi ∈ {0, 1}
P {Xi = 1} = p
P {Xi = 0} = q = 1 − p
2
µ = p
σ = pq
X1 + X2 + · · · + Xn
Xn =
Fraction of successes in n trials
n
k
P Xn =
n
= b (n, p , k ) =
n
k
p k q n−k
By CLT,
f Xn = x
(x − µ)2
∼ N µ,
exp
−
=q
n
2σ 2 n
2πσ 2 n
"
#
1
(x − p )2
= q
exp − 2pq n
2π pq n
σ2
1
"
#
Central Limit Theorem (CLT) : Bernoulli Trials
Xi ∈ {0, 1}
P {Xi = 1} = p
P {Xi = 0} = q = 1 − p
2
µ = p
σ = pq
X + X + · · · + Xn
Xn = 1 2
Fraction of successes in n trials
n
k
P Xn =
n
= b (n, p , k ) =
n
k
p k q n−k
By CLT, for large n,
f Xn = x
(x − µ)2
=q
exp
−
∼ N µ,
n
2σ 2 n
2πσ 2 n
"
#
1
(x − p )2
= q
exp − 2pq n
2π pq n
σ2
1
"
#
Central Limit Theorem (CLT) : Bernoulli Trials
Xi ∈ {0, 1}
P {Xi = 1} = p
P {Xi = 0} = q = 1 − p
2
µ = p
σ = pq
X + X + · · · + Xn
Xn = 1 2
Fraction of successes in n trials
n
k
P Xn =
n
= b (n, p , k ) =
n
k
p k q n−k
By CLT, for large n,
f Xn = x
(x − µ)2
=q
exp
−
∼ N µ,
n
2σ 2 n
2πσ 2 n
"
#
1
(x − p )2
= q
exp − 2pq n
2π pq n
σ2
1
"
#
Exponential Distribution
(
λ e −λ t
f (t ) =
0
,
,
t ≥0
otherwise
1λ : Mean
1 λ 2 : Variance
I Arrival time of telephone calls
I Bus arrival times at a bus stop
I Inter nucleotide distancein DNA sequences
Exponential Distribution
(
λ e −λ t
f (t ) =
0
,
,
t ≥0
otherwise
1λ : Mean
1 λ 2 : Variance
I Arrival time of telephone calls
I Bus arrival times at a bus stop
I Inter nucleotide distancein DNA sequences
Exponential Distribution
(
λ e −λ t
f (t ) =
0
,
,
t ≥0
otherwise
1λ : Mean
1 λ 2 : Variance
I Arrival time of telephone calls
I Bus arrival times at a bus stop
I Inter nucleotide distancein DNA sequences
Exponential Distribution
(
λ e −λ t
f (t ) =
0
,
,
t ≥0
otherwise
1λ : Mean
1 λ 2 : Variance
I Arrival time of telephone calls
I Bus arrival times at a bus stop
I Inter nucleotide distance in DNA sequences
Poisson Distribution
P (X = k ) = e −λ
λk
k!
k = 0, 1, 2, 3, ..., ∞
λ : Mean and Variance
I No. of phone calls at an exchange over a fixed duration of time
I No. of printing errors in a book
Poisson and Exponential Distributions
Poisson theorem or Law of rare events :
λk
n
lim
p k q n−k = e −λ
k!
np=λ , n→∞ k
Markov Models
I
Markov Process
I
I
I
I
Markov Property :
Current state depends stochastically only on the previous step
Ex: Random walk or Brownian motion
Non-Markovian processes may have Markovian representation
Hidden Markov Model
I
Markov process with unobserved (hidden) states
Markov Models
I
Markov Process
I
I
I
I
Markov Property :
Current state depends stochastically only on the previous step
Ex: Random walk or Brownian motion
Non-Markovian processes may have Markovian representation
Hidden Markov Model
I
Markov process with unobserved (hidden) states
Markov Models
I
Markov Process
I
I
I
I
Markov Property :
Current state depends stochastically only on the previous step
Ex: Random walk or Brownian motion
Non-Markovian processes may have Markovian representation
Hidden Markov Model
I
Markov process with unobserved (hidden) states
Markov Models
I
Markov Process
I
I
I
I
Markov Property :
Current state depends stochastically only on the previous step
Ex: Random walk or Brownian motion
Non-Markovian processes may have Markovian representation
Hidden Markov Model
I
Markov process with unobserved (hidden) states
Markov Models
I
Markov Process
I
I
I
I
Markov Property :
Current state depends stochastically only on the previous step
Ex: Random walk or Brownian motion
Non-Markovian processes may have Markovian representation
Hidden Markov Model
I
Markov process with unobserved (hidden) states
Markov Models
I
Markov Process
I
I
I
I
Markov Property :
Current state depends stochastically only on the previous step
Ex: Random walk or Brownian motion
Non-Markovian processes may have Markovian representation
Hidden Markov Model
I
Markov process with unobserved (hidden) states
Download