Random Processes

advertisement
PART II: RANDOM SIGNAL PROCESSING
Chapter 7 Random Processes
7.1 Introduction
- What is a random process? - The probability model used to characterize a random
signal is called a random process (or a stochastic process or a time series)
- Examples of random processes:
- a piece of sound collected in a train station (continuous)
- a section of facsimile in a fax machine (discrete)
- ……
- Many random processes are the outputs of linear systems, such as:
- the temperature in an industry oven
- the vibration of an electrical motor
- …
- Study random processes can help us understand the behavior of engineering systems
and hence to optimize their operations
- The road ahead:
- Chapter 7 = Chapter 3 in the textbook: Introduction to random processes
- Chapter 8 = Chapter 4 in the textbook: Linear systems
- Chapter 10 = Chapter 6 in the textbook: Signal Detection
- Chapter 11 = Chapter 7 in the textbook: Filtering
7.2 Definitions of Random Processes
- A random process can be described by X(t, ), where, t is time,  is a set of variables
can characterize the process and is the outcomes of random experiments E, and X
represents the waveform or the signal.
- The features of random variables:
- X(t, ) = {X(t, i) / i  S} = {x1(t), x2(t) …} is a collection of functions of time
- X(t0, i) = xi(t0) is a numerical value of the ith element at time t0.
- X(t, i) = xi(t) is a specific member function or deterministic function of time
- X(t0, ) = {X(t0, i) / i  S} = {x1(t0), x2(t0) …} is a collection of the numerical
values of the member functions at t = t0,
- Let us consider a simply example: a rotary machine has two states: normal and
abnormal. Under the normal condition, denoted as N, its vibration can be described by
the following function:
xn(t) = sint
Note that the angular speed is  = 1. On the other hand, under the abnormal
condition, denoted by A, its vibration is characterized by:
xa(t) = 2sint
Then, the measured vibration signal is a random process:
X(t, ) = {X(t, i) / i  {N, A}} = {xn(t), xa(t)} = {sint, 2sint}
In particular:
X(t, N) = sint
X(t, A) = 2sint
X(0, ) = {0, 0}
X(/2, ) = {1, 2}
X(0, N) = 0;
X(0, A) = 0;
X(/2, N) = 1;
X(/2, N) = 2;
-
Note that at each specific time, the random process becomes a random variable.
Hence, we can only calculate its probability. For example, if the machine is equally
likely to be in the state of normal and abnormal, then:
P(X(/2) = 1) = ½
-
This is called the probabilistic structure of the random process.
Similar to the (inform) definition of probability, let n be the total number trails at t =
t0, and k be the number of trail that result in X(t0, ) = a, then:
k
n  n
P( X (t 0 ,  )  a )  lim
-
We can define the joint and conditional probability in similar pattern.
The formal definition of random processes is based on the probability: Let S be the
sample space of a random experiment and t be a variable that can have values in the
set   R1 (the real line). A real-valued random process X(t), t  , is then a
measurable function on S that maps   S onto R1. It can be described by its (nth
order) distribution function:
FX (t1 ), X (t2 ),..., X (tn ) ( x1 , x2 ,..., xn )  P X (t1 )  x1 ,..., X (t n )  xn , for all n and t
-
Random processes can be classified according the characteristics of t and the random
variable X(t) at time t:
X(t) is continuous
t is continuous
Continuous random process
X(t) is discrete
Discrete random process
t is discrete
Continuous random
sequence
Discrete random sequence
-
Random processes can also be classified based on the probabilistic structure. If the
probability distribution of a random process does not depend on the time, t, then, the
process is called stationary. Otherwise, it is called non-stationary. That is:
if P(X(t)  x(t)) = P(X  x)
-
then X(t) is stationary
else X(t) is nonstationary
This will be further discussed in the subsequent sections.
Another important type of random processes is the Markov process. Given a random
process, X(t), if
PX (t k ) / X (t k 1 ),..., X (t1 )  PX (t k ) / X (t k 1 )
-
Then, it is a Markov process. The most important feature of a Markov process is that
its current probability depends on only the immediate past.
For the example above, the random process is continuous random process, it is
nonstationary but Markov.
7.3 Representation of Random Processes
- Random processes can be described by their probability structures. There are other
alternative ways to describe random processes and these descriptions may be easier to
use when solving certain problems.
- The first alternative is the joint probability distribution. In general:
- The first order: P( X (t1 )  a1 ) represents the instantaneous amplitude
- The second order: P( X (t1 )  a1 , X (t 2 )  a2 ) represents the structure of the signal
(which is related to the spectrum discussed in subsequent sections)
- The higher (kth) order: P( X (t1 )  a1 , X (t 2 )  a2 ,..., X (t k )  ak ) represents the
details of the signal.
-
The joint probability representation is easy to use. However, two different process
may have the same joint probability distributions.
The other alternative form is the analytical form. For example:
X(t) = Asin(t + )
-
where, A conforms a Normal distribution and  conforms an uniform distribution. In
general, the analytical form representation is unique but it is not always possible to
find an analytical form.
When two or more random processes are involved, say X(t) and Y(t), the following
concepts are important:
- The joint distribution:
P[X(t1) < x1, …, X(tn) < xn, Y(t’1) < y1, …, Y(t’n) < xn]
-
The cross-correlation function:

RXY (t1, t 2 )  EX * (t1 )Y (t 2 )
-
The cross-covariance function:

C XY (t1 , t 2 )  RXY (t1 , t 2 )   * X (t1 ) Y (t 2 )
-
The correlation coefficient:

rXY (t1 , t 2 ) 
-
C XY (t1 , t 2 )
C XX (t1 , t 2 )C XY (t1 , t 2 )
Equality: X(t) and Y(t) are equal if their respective member functions are identical
for each outcome of   S.
Uncorrelated: X(t) and Y(t) are uncorrelated if C XY (t1 , t2 ) = 0, t1, t2  
Orthogonal; X(t) and Y(t) are orthogonal if RXY (t1 , t 2 ) = 0, t1, t2  
Independent: X(t) and Y(t) are independent if
P[X(t1) < x1, …, X(tn) < xn, Y(t’1) < y1, …, Y(t’n) < xn] =
P[X(t1) < x1, …, X(tn) < xn] P[Y(t’1) < y1, …, Y(t’n) < xn], t1, t2 … t’1, t’2  
7.4 Some Special Random Processes
(1) Gaussian processes
- The most commonly used random process is Gaussian process. A Gaussian process,
X(t), t  , can be described by its probability density function:

f x (x)  2 
-
n2


1 2 1
 1

exp  x   x  1 x   x 
2


where, t1, t2, …, tn  , x = {x1, x2, …, xn}, and xi = X(ti). In addition, x is the mean
vector and  is the covariance matrix.
Gaussian processes have several important features: Let
 x1 
 11  12
x 

 2
 21  22
 ...   x 
 ...
...
x 
x      1  , x   1  ,   
...
 x2 
 xk  x 2 
 ...
 ... 
 ...
...
 

 xn 
 n1  n 2
Then:
...  1k
...  2 k
..
...
...  kk
...
...
...  nk
...  1n 
...  2 n 
... ...   11 12 

... ...   21  22 
... ... 

...  nn 
(a) Independence: if ij = 0, i  j, then X is independent.
(b) Partition at k: X1 has a k-dimensional multivariate Gaussian distribution with a
mean  x1 and covariance 11.
(c) Reduction: if A is a k  n matrix of rank k, then Y = AX has a k-variate Gaussian
distribution:
 y  A x
 y  A x AT
(d) Conditional probability: the conditional probability density function of X1 given
X2 = x2 is a k-dimensional multivariate Gaussian distribution with the mean and
covariance defined below:
X
1
X2

 EX 1 / X 2  x2   x1  12 221 x 2   x2

 X1 / X 2  11  12 221 21
-
An example: a work station sends out either a signal (1) or no signal (0). Because of
the noise disturbance, it is found that the signal can be described by a Gaussian
random process:
X(t) = {x1(t), x2(t)}
where, x1(t) ~ N(0, 1), x2(t) ~ N(1,1) and:
1 1


0 1
Note that X is not independent. The conditional probability of X1 given X2 = 2 is
Gaussian with:
X
1
X2
= 0 + (1)(1)-1(2 – 1) = 1
 X 1 / X 2 = 1 – (1)(1)-1(0) = 1
By the way, since the probability function is independent on the time, the signal is
stationary. Is it also Markov? (Yes)
(2) Another important type of random process is binary random process (random binary
waveform). It is often used in data communication systems.
- Illustration of a binary random process
1
T
D
t
-1
-
-
The properties of the binary random process:
- Each pulse has a rectangular shape with a fixed duration T and a random
amplitude 1.
- Pulse amplitude are equally likely to be 1 (can be converted to 0 and 1)
- All pulse amplitude are statistically independent
- The start time of the pulse sequences is arbitrary and is equally likely between 0
and T.
The mathematical expression:
X (t ) 

 Ak p(t  kT  D)
k  
-
where, p(t) is a unit amplitude pulse of duration T, Ak is a binary random variable
representing the kth pulse, and D is the random start time with uniform distribution in
the interval [0, T].
An example of binary random process, in the above figure: {…1, -1, 1, -1, -1, …}
The mathematical expectations:
- E{X(t)} = 0;
- E{X2(t)} = 1;
t t
- R XX (t1 , t 2 )  1  1 2
T
3.5. Stationarity
(1) Stationarity is one of the most important concepts in random signal processing. There
are two kinds of stationarities: strict-sense stationarity (SSS) and wide-sense
stationarity (WSS).
(2) Strict-sense stationarity
- definition: a random process X(t) is SSS if for all t1, t2, …, tk, t1 + , t2 + , …, tk + 
 and k = 1, 2, …,
P[X1(t)x1, X2(t)x2, …, Xk(t)xk] = P[X1(t+)x1, X2(t+)x2, …, Xk(t+)xk]
-
If the foregoing definition holds for all kth order distribution function k = 1, 2, …, N
but necessary for k > N, then the process is said to be Nth order stationary.
As a special case, the first order SSS is:
P[X(t)x] = P[X(t+)x]
-
That is, the probability distribution is independent of time t.
For a first order SSS random process:
E{X(t)} = x (a constant)
Note that the reverse is not true in general. That is, even if this equation holds, the
corresponding random process may not be SSS.
(3) Wide-sense stationarity
- WSS is much more widely used
- Definition: a random process X(t) is SSS if the following conditions are hold:
E{X(t)} = x (a constant)
E{X*(t)X(t+)} = Rxx()
-
where, X*(t) is the conjugate complex of X(t).
Example 1: Is the binary random process WSS? From (1), we know that:
E{X(t)} = 0;
R XX (t1 , t 2 )  1 
t1  t 2
T
Let t2 = t1 + , we have:
R XX (t1 , t1   )  1 
-
t1  t1  

 1
T
T
therefore, it is WSS.
multivariate random processes can be defined in a similar manner.
3.6. Autocorrelation and Power Spectrum
(1) Autocorrelation and power spectrum are very important for signal processing. It is
interesting to know that there are different ways to introduce power spectrum.
However, they all result in the same thing.
(2) Autocorrelation function
- the autocorrelation function of a real-valued WSS random process is defined as
RXX() = E{X(t)X(t+)}
-
the basic properties:
- it is related to the energy:
Rxx() = E{X2(t)} = average power (note that Rxx()  0)
-
it is an even function of :
Rxx() = Rxx(-)
it is bounded by Rxx() :
Rxx()  Rxx()
If X(t) contains a periodic component, then Rxx() will also contain a periodic
component
If lim R XX ( )  C , then C   X2
 
If Rxx() = Rxx(0) for some T0  0, then Rxx is periodic with a period T0
If Rxx(0) < 
(3) Cross-correlation function and its properties
- Cross-correlation function is defined as follows:
-
RXY() = E{X(t)Y(t+)}
-
The cross-correlation function has the following properties
- RXY() = RXY(-)
- R XY ( )  R XX (0) RYY (0)
-
1
R XX (0)  RYY (0)
2
RXY() = 0, if the processes are orthogonal, and
RXY() = XY, if the processes are independent
R XY ( ) 
3.7. Power Spectrum
(1) Definition and properties
- Based on the autocorrelation function, the power spectrum (Wiener-Khinchine) can
be defined as follows:

S XX ( f )  F RXX ( )   RXX ( ) exp(  j 2f )d

-
Note:
- This definition is applicable to all WSS random processes. How about the no
stationary random processes? (it does not make sense to calculate the power
spectrum).
- SXX(f) is called power spectral density function (psd).
Given the power spectral density function, SXX(f), the autocorrelation function can be
obtained:

RXX ( )  F 1 S XX ( f )   S XX ( f ) exp( j 2f )df

-
The properties of power spectrum:
- SXX(f) is real and nonnegative
-

The average power in X(t) is given by: E{ X 2 (t )}  RXX (0)   S XX ( f )df

- If X(t) real, RXX() will be even. Hence, SXX(f) will be even.
- If X(t) has periodic components, then SXX(f) will have impulses.
- We can define the cross-power spectral density function in a similar manner and find
its properties
(3) Lowpass and bandpass processes
- Definition: if the psd of a random process is zero for f> B (the bandwidth of the
process), then the process is called a lowpass process with the bandwidth B.
Similarly, the bandpass process is defined.
-B
-
bandpass
SXX(f)
lowpass
B
fc
fc+B/2
The power of a given frequency band, 0  f1 < f2, is as follows:
f2
PX [ f1 , f 2 ]  2  S XX ( f )df
f1
-
Some random processes may have psd functions with nonzero but insignificant values
over a large frequency bands, hence, the effective bandwidth (Beff) is defined:

Beff
1  S XX ( f )df
 
2 max S XX ( f )
This is illustrated as follows:
Max[SXX(f)]
Equal area
Beff
(4) Power spectral density function of random sequences
- We can define the power spectral function for random sequences by means of
summation (instead of integration).
(5) Examples.
Download